var/home/core/zuul-output/0000755000175000017500000000000015114057164014531 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114074745015501 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005077041115114074736017712 0ustar rootrootDec 03 16:03:40 crc systemd[1]: Starting Kubernetes Kubelet... Dec 03 16:03:40 crc restorecon[4695]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:40 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 16:03:41 crc restorecon[4695]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 16:03:41 crc restorecon[4695]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 03 16:03:41 crc kubenswrapper[4998]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 16:03:41 crc kubenswrapper[4998]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 03 16:03:41 crc kubenswrapper[4998]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 16:03:41 crc kubenswrapper[4998]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 16:03:41 crc kubenswrapper[4998]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 03 16:03:41 crc kubenswrapper[4998]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.496234 4998 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499306 4998 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499353 4998 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499362 4998 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499367 4998 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499372 4998 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499377 4998 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499381 4998 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499384 4998 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499388 4998 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499392 4998 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499396 4998 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499399 4998 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499404 4998 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499409 4998 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499413 4998 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499418 4998 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499424 4998 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499429 4998 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499434 4998 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499440 4998 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499446 4998 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499451 4998 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499456 4998 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499460 4998 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499464 4998 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499469 4998 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499473 4998 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499478 4998 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499482 4998 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499487 4998 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499492 4998 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499496 4998 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499500 4998 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499504 4998 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499507 4998 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499511 4998 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499515 4998 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499520 4998 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499524 4998 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499528 4998 feature_gate.go:330] unrecognized feature gate: Example Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499533 4998 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499537 4998 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499542 4998 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499546 4998 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499550 4998 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499554 4998 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499557 4998 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499561 4998 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499565 4998 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499569 4998 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499574 4998 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499578 4998 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499581 4998 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499585 4998 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499588 4998 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499592 4998 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499596 4998 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499599 4998 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499604 4998 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499607 4998 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499610 4998 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499614 4998 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499618 4998 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499621 4998 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499624 4998 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499628 4998 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499631 4998 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499635 4998 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499639 4998 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499642 4998 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.499645 4998 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499736 4998 flags.go:64] FLAG: --address="0.0.0.0" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499746 4998 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499771 4998 flags.go:64] FLAG: --anonymous-auth="true" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499783 4998 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499790 4998 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499795 4998 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499800 4998 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499805 4998 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499810 4998 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499814 4998 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499821 4998 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499826 4998 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499830 4998 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499834 4998 flags.go:64] FLAG: --cgroup-root="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499838 4998 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499843 4998 flags.go:64] FLAG: --client-ca-file="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499846 4998 flags.go:64] FLAG: --cloud-config="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499850 4998 flags.go:64] FLAG: --cloud-provider="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499854 4998 flags.go:64] FLAG: --cluster-dns="[]" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499865 4998 flags.go:64] FLAG: --cluster-domain="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499869 4998 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499873 4998 flags.go:64] FLAG: --config-dir="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499878 4998 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499883 4998 flags.go:64] FLAG: --container-log-max-files="5" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499889 4998 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499894 4998 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499899 4998 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499903 4998 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499908 4998 flags.go:64] FLAG: --contention-profiling="false" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499913 4998 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499917 4998 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499922 4998 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499926 4998 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499931 4998 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499935 4998 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499939 4998 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499943 4998 flags.go:64] FLAG: --enable-load-reader="false" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499947 4998 flags.go:64] FLAG: --enable-server="true" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499951 4998 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499958 4998 flags.go:64] FLAG: --event-burst="100" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499963 4998 flags.go:64] FLAG: --event-qps="50" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499967 4998 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499973 4998 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499977 4998 flags.go:64] FLAG: --eviction-hard="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499983 4998 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499987 4998 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499991 4998 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.499996 4998 flags.go:64] FLAG: --eviction-soft="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500000 4998 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500004 4998 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500008 4998 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500012 4998 flags.go:64] FLAG: --experimental-mounter-path="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500017 4998 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500021 4998 flags.go:64] FLAG: --fail-swap-on="true" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500026 4998 flags.go:64] FLAG: --feature-gates="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500030 4998 flags.go:64] FLAG: --file-check-frequency="20s" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500034 4998 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500039 4998 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500043 4998 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500047 4998 flags.go:64] FLAG: --healthz-port="10248" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500051 4998 flags.go:64] FLAG: --help="false" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500056 4998 flags.go:64] FLAG: --hostname-override="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500060 4998 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500064 4998 flags.go:64] FLAG: --http-check-frequency="20s" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500068 4998 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500072 4998 flags.go:64] FLAG: --image-credential-provider-config="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500076 4998 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500080 4998 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500084 4998 flags.go:64] FLAG: --image-service-endpoint="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500088 4998 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500092 4998 flags.go:64] FLAG: --kube-api-burst="100" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500096 4998 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500100 4998 flags.go:64] FLAG: --kube-api-qps="50" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500104 4998 flags.go:64] FLAG: --kube-reserved="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500108 4998 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500113 4998 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500117 4998 flags.go:64] FLAG: --kubelet-cgroups="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500121 4998 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500125 4998 flags.go:64] FLAG: --lock-file="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500128 4998 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500133 4998 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500137 4998 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500144 4998 flags.go:64] FLAG: --log-json-split-stream="false" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500148 4998 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500152 4998 flags.go:64] FLAG: --log-text-split-stream="false" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500157 4998 flags.go:64] FLAG: --logging-format="text" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500162 4998 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500166 4998 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500170 4998 flags.go:64] FLAG: --manifest-url="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500174 4998 flags.go:64] FLAG: --manifest-url-header="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500180 4998 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500184 4998 flags.go:64] FLAG: --max-open-files="1000000" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500189 4998 flags.go:64] FLAG: --max-pods="110" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500193 4998 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500197 4998 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500201 4998 flags.go:64] FLAG: --memory-manager-policy="None" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500205 4998 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500209 4998 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500213 4998 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500217 4998 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500226 4998 flags.go:64] FLAG: --node-status-max-images="50" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500230 4998 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500234 4998 flags.go:64] FLAG: --oom-score-adj="-999" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500238 4998 flags.go:64] FLAG: --pod-cidr="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500242 4998 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500248 4998 flags.go:64] FLAG: --pod-manifest-path="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500252 4998 flags.go:64] FLAG: --pod-max-pids="-1" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500257 4998 flags.go:64] FLAG: --pods-per-core="0" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500261 4998 flags.go:64] FLAG: --port="10250" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500265 4998 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500269 4998 flags.go:64] FLAG: --provider-id="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500274 4998 flags.go:64] FLAG: --qos-reserved="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500278 4998 flags.go:64] FLAG: --read-only-port="10255" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500282 4998 flags.go:64] FLAG: --register-node="true" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500287 4998 flags.go:64] FLAG: --register-schedulable="true" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500291 4998 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500297 4998 flags.go:64] FLAG: --registry-burst="10" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500301 4998 flags.go:64] FLAG: --registry-qps="5" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500305 4998 flags.go:64] FLAG: --reserved-cpus="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500309 4998 flags.go:64] FLAG: --reserved-memory="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500314 4998 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500319 4998 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500323 4998 flags.go:64] FLAG: --rotate-certificates="false" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500327 4998 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500331 4998 flags.go:64] FLAG: --runonce="false" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500335 4998 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500339 4998 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500343 4998 flags.go:64] FLAG: --seccomp-default="false" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500347 4998 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500351 4998 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500355 4998 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500359 4998 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500364 4998 flags.go:64] FLAG: --storage-driver-password="root" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500368 4998 flags.go:64] FLAG: --storage-driver-secure="false" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500372 4998 flags.go:64] FLAG: --storage-driver-table="stats" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500376 4998 flags.go:64] FLAG: --storage-driver-user="root" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500380 4998 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500384 4998 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500389 4998 flags.go:64] FLAG: --system-cgroups="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500394 4998 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500401 4998 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500407 4998 flags.go:64] FLAG: --tls-cert-file="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500411 4998 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500417 4998 flags.go:64] FLAG: --tls-min-version="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500421 4998 flags.go:64] FLAG: --tls-private-key-file="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500426 4998 flags.go:64] FLAG: --topology-manager-policy="none" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500431 4998 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500436 4998 flags.go:64] FLAG: --topology-manager-scope="container" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500440 4998 flags.go:64] FLAG: --v="2" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500446 4998 flags.go:64] FLAG: --version="false" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500451 4998 flags.go:64] FLAG: --vmodule="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500456 4998 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500460 4998 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500559 4998 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500564 4998 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500568 4998 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500572 4998 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500576 4998 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500580 4998 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500583 4998 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500587 4998 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500590 4998 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500594 4998 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500597 4998 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500601 4998 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500604 4998 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500608 4998 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500611 4998 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500615 4998 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500618 4998 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500622 4998 feature_gate.go:330] unrecognized feature gate: Example Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500633 4998 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500637 4998 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500640 4998 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500644 4998 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500648 4998 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500653 4998 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500657 4998 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500662 4998 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500666 4998 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500670 4998 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500673 4998 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500677 4998 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500681 4998 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500685 4998 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500689 4998 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500693 4998 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500697 4998 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500702 4998 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500706 4998 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500709 4998 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500713 4998 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500717 4998 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500720 4998 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500723 4998 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500729 4998 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500733 4998 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500738 4998 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500742 4998 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500745 4998 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500762 4998 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500766 4998 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500770 4998 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500775 4998 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500779 4998 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500782 4998 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500786 4998 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500789 4998 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500795 4998 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500798 4998 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500802 4998 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500805 4998 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500808 4998 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500812 4998 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500815 4998 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500819 4998 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500822 4998 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500826 4998 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500829 4998 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500833 4998 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500836 4998 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500840 4998 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500843 4998 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.500847 4998 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.500986 4998 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.514355 4998 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.514411 4998 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.514646 4998 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.514674 4998 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.514909 4998 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.514925 4998 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.514937 4998 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.514947 4998 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.514957 4998 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.514967 4998 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.514976 4998 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.514987 4998 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.514995 4998 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515004 4998 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515014 4998 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515023 4998 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515031 4998 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515040 4998 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515049 4998 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515059 4998 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515069 4998 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515079 4998 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515088 4998 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515112 4998 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515124 4998 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515135 4998 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515145 4998 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515154 4998 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515162 4998 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515170 4998 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515178 4998 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515186 4998 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515195 4998 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515204 4998 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515212 4998 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515219 4998 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515228 4998 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515236 4998 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515245 4998 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515253 4998 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515261 4998 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515270 4998 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515278 4998 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515287 4998 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515295 4998 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515303 4998 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515312 4998 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515323 4998 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515338 4998 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515349 4998 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515359 4998 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515370 4998 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515379 4998 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515388 4998 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515397 4998 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515405 4998 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515414 4998 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515422 4998 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515430 4998 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515452 4998 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515461 4998 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515472 4998 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515484 4998 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515493 4998 feature_gate.go:330] unrecognized feature gate: Example Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515501 4998 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515510 4998 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515520 4998 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515529 4998 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515538 4998 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515546 4998 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515558 4998 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515569 4998 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515581 4998 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.515597 4998 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515965 4998 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515983 4998 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.515993 4998 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516002 4998 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516011 4998 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516020 4998 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516028 4998 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516037 4998 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516046 4998 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516055 4998 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516063 4998 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516071 4998 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516079 4998 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516091 4998 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516104 4998 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516113 4998 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516122 4998 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516131 4998 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516139 4998 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516149 4998 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516157 4998 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516191 4998 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516200 4998 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516209 4998 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516218 4998 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516226 4998 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516234 4998 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516243 4998 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516251 4998 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516261 4998 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516270 4998 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516279 4998 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516289 4998 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516298 4998 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516307 4998 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516316 4998 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516324 4998 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516332 4998 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516340 4998 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516348 4998 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516357 4998 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516365 4998 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516373 4998 feature_gate.go:330] unrecognized feature gate: Example Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516381 4998 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516389 4998 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516398 4998 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516406 4998 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516415 4998 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516424 4998 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516432 4998 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516441 4998 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516450 4998 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516458 4998 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516467 4998 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516475 4998 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516484 4998 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516492 4998 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516518 4998 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516528 4998 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516538 4998 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516546 4998 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516555 4998 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516563 4998 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516572 4998 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516583 4998 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516597 4998 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516608 4998 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516618 4998 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516628 4998 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516639 4998 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.516650 4998 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.516664 4998 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.517305 4998 server.go:940] "Client rotation is on, will bootstrap in background" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.524705 4998 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.524916 4998 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.525888 4998 server.go:997] "Starting client certificate rotation" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.525931 4998 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.526413 4998 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-05 08:39:07.212082431 +0000 UTC Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.526548 4998 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 784h35m25.685540676s for next certificate rotation Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.534201 4998 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.536885 4998 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.547226 4998 log.go:25] "Validated CRI v1 runtime API" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.572204 4998 log.go:25] "Validated CRI v1 image API" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.574300 4998 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.577134 4998 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-03-15-58-49-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.577169 4998 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.591043 4998 manager.go:217] Machine: {Timestamp:2025-12-03 16:03:41.589991079 +0000 UTC m=+0.201691322 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:0d8400b0-b239-4c8c-8060-4c5edd34a62b BootID:ef69ca4e-e477-4474-a354-15f3061d008d Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:df:55:33 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:df:55:33 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:63:51:5a Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:e1:48:c6 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:77:f6:2d Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:d0:c1:0f Speed:-1 Mtu:1496} {Name:eth10 MacAddress:d2:e5:1b:60:5a:bb Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:de:a5:8e:6e:b9:80 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.591238 4998 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.591422 4998 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.591846 4998 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.591994 4998 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.592028 4998 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.592216 4998 topology_manager.go:138] "Creating topology manager with none policy" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.592227 4998 container_manager_linux.go:303] "Creating device plugin manager" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.592384 4998 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.592421 4998 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.592595 4998 state_mem.go:36] "Initialized new in-memory state store" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.592678 4998 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.593666 4998 kubelet.go:418] "Attempting to sync node with API server" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.593690 4998 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.593719 4998 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.593734 4998 kubelet.go:324] "Adding apiserver pod source" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.593764 4998 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.598569 4998 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.64:6443: connect: connection refused Dec 03 16:03:41 crc kubenswrapper[4998]: E1203 16:03:41.598715 4998 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.64:6443: connect: connection refused" logger="UnhandledError" Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.598844 4998 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.64:6443: connect: connection refused Dec 03 16:03:41 crc kubenswrapper[4998]: E1203 16:03:41.599700 4998 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.64:6443: connect: connection refused" logger="UnhandledError" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.599808 4998 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.600365 4998 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.601460 4998 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.602387 4998 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.602431 4998 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.602448 4998 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.602462 4998 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.602488 4998 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.602502 4998 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.602518 4998 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.602544 4998 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.602562 4998 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.602576 4998 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.602596 4998 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.602611 4998 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.603419 4998 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.604247 4998 server.go:1280] "Started kubelet" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.604571 4998 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.604674 4998 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.605355 4998 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.64:6443: connect: connection refused Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.605570 4998 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.606718 4998 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.606806 4998 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.606940 4998 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.606968 4998 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.606900 4998 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 08:27:06.471711822 +0000 UTC Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.607013 4998 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 520h23m24.864707259s for next certificate rotation Dec 03 16:03:41 crc kubenswrapper[4998]: E1203 16:03:41.607113 4998 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.607228 4998 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 03 16:03:41 crc systemd[1]: Started Kubernetes Kubelet. Dec 03 16:03:41 crc kubenswrapper[4998]: E1203 16:03:41.608931 4998 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" interval="200ms" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.609365 4998 server.go:460] "Adding debug handlers to kubelet server" Dec 03 16:03:41 crc kubenswrapper[4998]: E1203 16:03:41.608984 4998 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.64:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187dc0141c3490d7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 16:03:41.604188375 +0000 UTC m=+0.215888638,LastTimestamp:2025-12-03 16:03:41.604188375 +0000 UTC m=+0.215888638,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.609510 4998 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.64:6443: connect: connection refused Dec 03 16:03:41 crc kubenswrapper[4998]: E1203 16:03:41.609615 4998 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.64:6443: connect: connection refused" logger="UnhandledError" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.609852 4998 factory.go:55] Registering systemd factory Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.609873 4998 factory.go:221] Registration of the systemd container factory successfully Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.610284 4998 factory.go:153] Registering CRI-O factory Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.610308 4998 factory.go:221] Registration of the crio container factory successfully Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.610409 4998 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.610444 4998 factory.go:103] Registering Raw factory Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.610464 4998 manager.go:1196] Started watching for new ooms in manager Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.611220 4998 manager.go:319] Starting recovery of all containers Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.635490 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.635655 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.635701 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.635744 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.635822 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.635852 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.635880 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.635907 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.635938 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.635963 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.635992 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.636018 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.636045 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.636077 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.636103 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.636129 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.636201 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.636231 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.636258 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.636286 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.636313 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.636342 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.636370 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.636398 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.636424 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.636452 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.636493 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.636524 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.636550 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.636580 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.636611 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.636672 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.636835 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.636911 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.636951 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.636988 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637015 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637049 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637075 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637100 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637174 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637202 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637222 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637242 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637271 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637291 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637317 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637338 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637362 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637382 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637402 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637421 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637450 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637491 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637516 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637594 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637630 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637656 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637683 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637708 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637732 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.637786 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.640523 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.640599 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.640639 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.640672 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.640700 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.640726 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.640785 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.640817 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.640848 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.640876 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.640906 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.640933 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.640962 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.640992 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641023 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641053 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641078 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641159 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641195 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641221 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641247 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641274 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641301 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641328 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641355 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641380 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641408 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641437 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641466 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641491 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641518 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641637 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641652 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641664 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641674 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641699 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641711 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641723 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641734 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641743 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641792 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641804 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641824 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641840 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641894 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641907 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641919 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641950 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641964 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641976 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641986 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.641998 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642024 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642035 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642047 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642057 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642067 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642077 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642103 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642115 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642127 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642139 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642149 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642175 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642189 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642200 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642211 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642222 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642232 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642259 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642271 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642282 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642294 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642305 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642333 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642344 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642356 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642366 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642377 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642403 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642413 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642425 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642436 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642447 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642458 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642484 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642495 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642506 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642517 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642527 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642538 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642564 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642577 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642588 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642599 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642611 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642638 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642650 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642666 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.642679 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643281 4998 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643305 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643319 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643346 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643359 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643369 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643380 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643391 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643401 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643429 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643440 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643453 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643464 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643509 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643523 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643535 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643547 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643557 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643584 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643595 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643635 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643668 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643680 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643690 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643700 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643711 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643721 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643745 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643767 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643780 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643795 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643822 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643833 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643847 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643858 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643871 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643897 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643909 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643919 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643931 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643942 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643958 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643985 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.643996 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.644008 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.644018 4998 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.644028 4998 reconstruct.go:97] "Volume reconstruction finished" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.644036 4998 reconciler.go:26] "Reconciler: start to sync state" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.653877 4998 manager.go:324] Recovery completed Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.669121 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.671578 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.671672 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.671697 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.673590 4998 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.673727 4998 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.673864 4998 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.673923 4998 state_mem.go:36] "Initialized new in-memory state store" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.676425 4998 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.676473 4998 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.676506 4998 kubelet.go:2335] "Starting kubelet main sync loop" Dec 03 16:03:41 crc kubenswrapper[4998]: E1203 16:03:41.676551 4998 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 03 16:03:41 crc kubenswrapper[4998]: W1203 16:03:41.677710 4998 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.64:6443: connect: connection refused Dec 03 16:03:41 crc kubenswrapper[4998]: E1203 16:03:41.677822 4998 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.64:6443: connect: connection refused" logger="UnhandledError" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.683421 4998 policy_none.go:49] "None policy: Start" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.685090 4998 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.685140 4998 state_mem.go:35] "Initializing new in-memory state store" Dec 03 16:03:41 crc kubenswrapper[4998]: E1203 16:03:41.708178 4998 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.736647 4998 manager.go:334] "Starting Device Plugin manager" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.736851 4998 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.736932 4998 server.go:79] "Starting device plugin registration server" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.737706 4998 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.737806 4998 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.738951 4998 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.739047 4998 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.739056 4998 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 03 16:03:41 crc kubenswrapper[4998]: E1203 16:03:41.750988 4998 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.777127 4998 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.777340 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.780064 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.780127 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.780145 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.780338 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.780529 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.780651 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.781525 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.781617 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.781647 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.782134 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.782312 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.782389 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.782945 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.783001 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.783013 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.783499 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.783551 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.783563 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.783733 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.783931 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.783995 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.785029 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.785073 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.785091 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.785437 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.785472 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.785485 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.785603 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.785626 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.785699 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.785731 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.785672 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.785806 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.786359 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.786387 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.786397 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.786547 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.786573 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.786610 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.786634 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.786647 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.787521 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.787548 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.787558 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:41 crc kubenswrapper[4998]: E1203 16:03:41.811320 4998 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" interval="400ms" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.839053 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.840469 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.840574 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.840650 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.840724 4998 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 16:03:41 crc kubenswrapper[4998]: E1203 16:03:41.841529 4998 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.64:6443: connect: connection refused" node="crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.846686 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.846734 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.846799 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.846964 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.847096 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.847328 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.847393 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.847427 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.847475 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.847526 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.847591 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.847737 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.847856 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.847945 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.848030 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.949330 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.949435 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.949490 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.949534 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.949616 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.949799 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.949886 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.949732 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.949616 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.949743 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.949849 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.950165 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.949683 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.950230 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.950283 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.950298 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.950333 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.950378 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.950425 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.950465 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.950516 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.950427 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.950554 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.950608 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.950556 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.950704 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.950805 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.950877 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.950946 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 16:03:41 crc kubenswrapper[4998]: I1203 16:03:41.951049 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.042530 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.046622 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.046698 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.046721 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.046815 4998 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 16:03:42 crc kubenswrapper[4998]: E1203 16:03:42.047550 4998 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.64:6443: connect: connection refused" node="crc" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.112131 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.132848 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.149429 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 16:03:42 crc kubenswrapper[4998]: W1203 16:03:42.153335 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-f82f3ecdb295658338d9e031e9cb1435fa1f8e1392a14619ff9b976f7fe2cc09 WatchSource:0}: Error finding container f82f3ecdb295658338d9e031e9cb1435fa1f8e1392a14619ff9b976f7fe2cc09: Status 404 returned error can't find the container with id f82f3ecdb295658338d9e031e9cb1435fa1f8e1392a14619ff9b976f7fe2cc09 Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.163036 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.165318 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:03:42 crc kubenswrapper[4998]: W1203 16:03:42.180127 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-5833c6c340a993da4069e47b0eda4f6f135e0fe457e75be2a08958a8375fb543 WatchSource:0}: Error finding container 5833c6c340a993da4069e47b0eda4f6f135e0fe457e75be2a08958a8375fb543: Status 404 returned error can't find the container with id 5833c6c340a993da4069e47b0eda4f6f135e0fe457e75be2a08958a8375fb543 Dec 03 16:03:42 crc kubenswrapper[4998]: W1203 16:03:42.183981 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-c935122fbde88c41fa71c0fed2711177e4b7963e43c675aa8ffccc79e1337e66 WatchSource:0}: Error finding container c935122fbde88c41fa71c0fed2711177e4b7963e43c675aa8ffccc79e1337e66: Status 404 returned error can't find the container with id c935122fbde88c41fa71c0fed2711177e4b7963e43c675aa8ffccc79e1337e66 Dec 03 16:03:42 crc kubenswrapper[4998]: W1203 16:03:42.197478 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-659f4556ccf1b9fa376fb0fb9141867f4b5916b47f1cbcb5974fa2a36909e599 WatchSource:0}: Error finding container 659f4556ccf1b9fa376fb0fb9141867f4b5916b47f1cbcb5974fa2a36909e599: Status 404 returned error can't find the container with id 659f4556ccf1b9fa376fb0fb9141867f4b5916b47f1cbcb5974fa2a36909e599 Dec 03 16:03:42 crc kubenswrapper[4998]: E1203 16:03:42.212968 4998 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" interval="800ms" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.448242 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.450444 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.450507 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.450526 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.450564 4998 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 16:03:42 crc kubenswrapper[4998]: E1203 16:03:42.451319 4998 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.64:6443: connect: connection refused" node="crc" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.606424 4998 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.64:6443: connect: connection refused Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.684022 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a"} Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.684143 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"659f4556ccf1b9fa376fb0fb9141867f4b5916b47f1cbcb5974fa2a36909e599"} Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.685516 4998 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4" exitCode=0 Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.685605 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4"} Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.685660 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c935122fbde88c41fa71c0fed2711177e4b7963e43c675aa8ffccc79e1337e66"} Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.685793 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.687305 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.687345 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.687357 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.689002 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.690104 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.690125 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.690132 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.692854 4998 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="8d25d7c896b5f131195161820371e78cbefd894d5aaf86bbd17a61ab7f617bb2" exitCode=0 Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.692915 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"8d25d7c896b5f131195161820371e78cbefd894d5aaf86bbd17a61ab7f617bb2"} Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.692936 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"5833c6c340a993da4069e47b0eda4f6f135e0fe457e75be2a08958a8375fb543"} Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.692993 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.693747 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.693802 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.693815 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.695139 4998 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034" exitCode=0 Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.695198 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034"} Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.695218 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"cb15fd2671e3bd4f8f6f3a360b00a7df55cd5040bc50c39a19e87d957c766ee7"} Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.695370 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.696376 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.696401 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.696409 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.698209 4998 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="92f6817f5eb33e35e106cb224f0fd8a31f3ff73d3a12621774680bbec5d24a2f" exitCode=0 Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.698231 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"92f6817f5eb33e35e106cb224f0fd8a31f3ff73d3a12621774680bbec5d24a2f"} Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.698243 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f82f3ecdb295658338d9e031e9cb1435fa1f8e1392a14619ff9b976f7fe2cc09"} Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.698317 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.705556 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.705585 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:42 crc kubenswrapper[4998]: I1203 16:03:42.705597 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:42 crc kubenswrapper[4998]: W1203 16:03:42.959394 4998 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.64:6443: connect: connection refused Dec 03 16:03:42 crc kubenswrapper[4998]: E1203 16:03:42.959508 4998 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.64:6443: connect: connection refused" logger="UnhandledError" Dec 03 16:03:42 crc kubenswrapper[4998]: W1203 16:03:42.991130 4998 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.64:6443: connect: connection refused Dec 03 16:03:42 crc kubenswrapper[4998]: E1203 16:03:42.991267 4998 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.64:6443: connect: connection refused" logger="UnhandledError" Dec 03 16:03:43 crc kubenswrapper[4998]: E1203 16:03:43.014396 4998 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" interval="1.6s" Dec 03 16:03:43 crc kubenswrapper[4998]: W1203 16:03:43.030147 4998 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.64:6443: connect: connection refused Dec 03 16:03:43 crc kubenswrapper[4998]: E1203 16:03:43.030263 4998 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.64:6443: connect: connection refused" logger="UnhandledError" Dec 03 16:03:43 crc kubenswrapper[4998]: W1203 16:03:43.204294 4998 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.64:6443: connect: connection refused Dec 03 16:03:43 crc kubenswrapper[4998]: E1203 16:03:43.204402 4998 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.64:6443: connect: connection refused" logger="UnhandledError" Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.252360 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.255438 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.255487 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.255499 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.255527 4998 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.709956 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a"} Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.710014 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2"} Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.710030 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb"} Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.710043 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079"} Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.712776 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"c9b7ae668ca10ea755cabe43c56545791b51cb31e3733764494bdb9f8daafe67"} Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.712963 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.714274 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.714315 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.714328 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.716249 4998 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d" exitCode=0 Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.716340 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d"} Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.716417 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.717890 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.717922 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.717934 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.721059 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"235459dff78aa6c80163b68bdb48c604268b9f8bb340326a52de53d44bcb98bf"} Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.721114 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"36b5dff655685383dcbeeb2a97f745bc3708c15cdfcf113ec1e2b230daf70bc5"} Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.721139 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ff6d272ee68c57986031e2d525eced78a6c7316a8980d1bae00590e14ce4fcdb"} Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.721234 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.722105 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.722134 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.722147 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.724519 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a9cdcf95d23503cce6d3fbc6de109783933aaaec4eefccd6f1834e234ed41f6e"} Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.724559 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c911081098fed822c164c7e59c1518ec11cf1c612e3ac4c84af8e694498c915d"} Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.724572 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a53614cedb63f775d2b563e4405055cd6702c3f35980bc092ac07a3c9ff69aca"} Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.724614 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.725689 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.725742 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:43 crc kubenswrapper[4998]: I1203 16:03:43.725792 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:44 crc kubenswrapper[4998]: I1203 16:03:44.731972 4998 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610" exitCode=0 Dec 03 16:03:44 crc kubenswrapper[4998]: I1203 16:03:44.732112 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610"} Dec 03 16:03:44 crc kubenswrapper[4998]: I1203 16:03:44.732171 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:44 crc kubenswrapper[4998]: I1203 16:03:44.733422 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:44 crc kubenswrapper[4998]: I1203 16:03:44.733474 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:44 crc kubenswrapper[4998]: I1203 16:03:44.733491 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:44 crc kubenswrapper[4998]: I1203 16:03:44.738262 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955"} Dec 03 16:03:44 crc kubenswrapper[4998]: I1203 16:03:44.738294 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:44 crc kubenswrapper[4998]: I1203 16:03:44.738338 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:44 crc kubenswrapper[4998]: I1203 16:03:44.740231 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:44 crc kubenswrapper[4998]: I1203 16:03:44.740237 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:44 crc kubenswrapper[4998]: I1203 16:03:44.740287 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:44 crc kubenswrapper[4998]: I1203 16:03:44.740357 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:44 crc kubenswrapper[4998]: I1203 16:03:44.740317 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:44 crc kubenswrapper[4998]: I1203 16:03:44.740427 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:45 crc kubenswrapper[4998]: I1203 16:03:45.238644 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:03:45 crc kubenswrapper[4998]: I1203 16:03:45.425763 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:03:45 crc kubenswrapper[4998]: I1203 16:03:45.505677 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:03:45 crc kubenswrapper[4998]: I1203 16:03:45.751263 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28"} Dec 03 16:03:45 crc kubenswrapper[4998]: I1203 16:03:45.752076 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c"} Dec 03 16:03:45 crc kubenswrapper[4998]: I1203 16:03:45.752121 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137"} Dec 03 16:03:45 crc kubenswrapper[4998]: I1203 16:03:45.751478 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:45 crc kubenswrapper[4998]: I1203 16:03:45.751409 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:45 crc kubenswrapper[4998]: I1203 16:03:45.753421 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:45 crc kubenswrapper[4998]: I1203 16:03:45.753479 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:45 crc kubenswrapper[4998]: I1203 16:03:45.753498 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:45 crc kubenswrapper[4998]: I1203 16:03:45.754148 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:45 crc kubenswrapper[4998]: I1203 16:03:45.754191 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:45 crc kubenswrapper[4998]: I1203 16:03:45.754204 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:46 crc kubenswrapper[4998]: I1203 16:03:46.156293 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:03:46 crc kubenswrapper[4998]: I1203 16:03:46.761269 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623"} Dec 03 16:03:46 crc kubenswrapper[4998]: I1203 16:03:46.761361 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b"} Dec 03 16:03:46 crc kubenswrapper[4998]: I1203 16:03:46.761403 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:46 crc kubenswrapper[4998]: I1203 16:03:46.761543 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:46 crc kubenswrapper[4998]: I1203 16:03:46.761552 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:46 crc kubenswrapper[4998]: I1203 16:03:46.763648 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:46 crc kubenswrapper[4998]: I1203 16:03:46.763698 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:46 crc kubenswrapper[4998]: I1203 16:03:46.763699 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:46 crc kubenswrapper[4998]: I1203 16:03:46.763716 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:46 crc kubenswrapper[4998]: I1203 16:03:46.763734 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:46 crc kubenswrapper[4998]: I1203 16:03:46.763750 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:46 crc kubenswrapper[4998]: I1203 16:03:46.763660 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:46 crc kubenswrapper[4998]: I1203 16:03:46.763826 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:46 crc kubenswrapper[4998]: I1203 16:03:46.763842 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:47 crc kubenswrapper[4998]: I1203 16:03:47.275033 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:03:47 crc kubenswrapper[4998]: I1203 16:03:47.360585 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:03:47 crc kubenswrapper[4998]: I1203 16:03:47.369976 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:03:47 crc kubenswrapper[4998]: I1203 16:03:47.765284 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:47 crc kubenswrapper[4998]: I1203 16:03:47.765314 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:47 crc kubenswrapper[4998]: I1203 16:03:47.765998 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:47 crc kubenswrapper[4998]: I1203 16:03:47.766948 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:47 crc kubenswrapper[4998]: I1203 16:03:47.766966 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:47 crc kubenswrapper[4998]: I1203 16:03:47.766995 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:47 crc kubenswrapper[4998]: I1203 16:03:47.767008 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:47 crc kubenswrapper[4998]: I1203 16:03:47.767029 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:47 crc kubenswrapper[4998]: I1203 16:03:47.767015 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:47 crc kubenswrapper[4998]: I1203 16:03:47.767581 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:47 crc kubenswrapper[4998]: I1203 16:03:47.767636 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:47 crc kubenswrapper[4998]: I1203 16:03:47.767660 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:48 crc kubenswrapper[4998]: I1203 16:03:48.236948 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 03 16:03:48 crc kubenswrapper[4998]: I1203 16:03:48.509976 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:03:48 crc kubenswrapper[4998]: I1203 16:03:48.769054 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:48 crc kubenswrapper[4998]: I1203 16:03:48.769153 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:48 crc kubenswrapper[4998]: I1203 16:03:48.769054 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:48 crc kubenswrapper[4998]: I1203 16:03:48.771089 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:48 crc kubenswrapper[4998]: I1203 16:03:48.771149 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:48 crc kubenswrapper[4998]: I1203 16:03:48.771148 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:48 crc kubenswrapper[4998]: I1203 16:03:48.771175 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:48 crc kubenswrapper[4998]: I1203 16:03:48.771197 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:48 crc kubenswrapper[4998]: I1203 16:03:48.771219 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:48 crc kubenswrapper[4998]: I1203 16:03:48.771357 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:48 crc kubenswrapper[4998]: I1203 16:03:48.771403 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:48 crc kubenswrapper[4998]: I1203 16:03:48.771419 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:49 crc kubenswrapper[4998]: I1203 16:03:49.157110 4998 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 16:03:49 crc kubenswrapper[4998]: I1203 16:03:49.157263 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 16:03:51 crc kubenswrapper[4998]: E1203 16:03:51.751235 4998 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 16:03:52 crc kubenswrapper[4998]: I1203 16:03:52.969333 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 03 16:03:52 crc kubenswrapper[4998]: I1203 16:03:52.970206 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:52 crc kubenswrapper[4998]: I1203 16:03:52.972413 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:52 crc kubenswrapper[4998]: I1203 16:03:52.972463 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:52 crc kubenswrapper[4998]: I1203 16:03:52.972481 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:52 crc kubenswrapper[4998]: I1203 16:03:52.984368 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 16:03:52 crc kubenswrapper[4998]: I1203 16:03:52.984604 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:52 crc kubenswrapper[4998]: I1203 16:03:52.986123 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:52 crc kubenswrapper[4998]: I1203 16:03:52.986184 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:52 crc kubenswrapper[4998]: I1203 16:03:52.986203 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:53 crc kubenswrapper[4998]: E1203 16:03:53.258865 4998 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 03 16:03:53 crc kubenswrapper[4998]: I1203 16:03:53.606525 4998 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 03 16:03:54 crc kubenswrapper[4998]: E1203 16:03:54.615105 4998 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 03 16:03:54 crc kubenswrapper[4998]: I1203 16:03:54.747033 4998 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 16:03:54 crc kubenswrapper[4998]: I1203 16:03:54.747095 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 16:03:54 crc kubenswrapper[4998]: I1203 16:03:54.753196 4998 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 16:03:54 crc kubenswrapper[4998]: I1203 16:03:54.753305 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 16:03:54 crc kubenswrapper[4998]: I1203 16:03:54.860039 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:54 crc kubenswrapper[4998]: I1203 16:03:54.861486 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:54 crc kubenswrapper[4998]: I1203 16:03:54.861517 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:54 crc kubenswrapper[4998]: I1203 16:03:54.861529 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:54 crc kubenswrapper[4998]: I1203 16:03:54.861552 4998 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 16:03:55 crc kubenswrapper[4998]: I1203 16:03:55.513946 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:03:55 crc kubenswrapper[4998]: I1203 16:03:55.514145 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:55 crc kubenswrapper[4998]: I1203 16:03:55.515655 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:55 crc kubenswrapper[4998]: I1203 16:03:55.515692 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:55 crc kubenswrapper[4998]: I1203 16:03:55.515701 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:57 crc kubenswrapper[4998]: I1203 16:03:57.280980 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:03:57 crc kubenswrapper[4998]: I1203 16:03:57.281180 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:57 crc kubenswrapper[4998]: I1203 16:03:57.282329 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:57 crc kubenswrapper[4998]: I1203 16:03:57.282360 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:57 crc kubenswrapper[4998]: I1203 16:03:57.282369 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:57 crc kubenswrapper[4998]: I1203 16:03:57.290246 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:03:57 crc kubenswrapper[4998]: I1203 16:03:57.799411 4998 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 16:03:57 crc kubenswrapper[4998]: I1203 16:03:57.799467 4998 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 16:03:57 crc kubenswrapper[4998]: I1203 16:03:57.800569 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:57 crc kubenswrapper[4998]: I1203 16:03:57.800620 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:57 crc kubenswrapper[4998]: I1203 16:03:57.800692 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.156543 4998 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.156624 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.731419 4998 trace.go:236] Trace[1120880201]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 16:03:45.786) (total time: 13944ms): Dec 03 16:03:59 crc kubenswrapper[4998]: Trace[1120880201]: ---"Objects listed" error: 13944ms (16:03:59.731) Dec 03 16:03:59 crc kubenswrapper[4998]: Trace[1120880201]: [13.944441069s] [13.944441069s] END Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.731450 4998 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.731955 4998 trace.go:236] Trace[1694827292]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 16:03:45.961) (total time: 13770ms): Dec 03 16:03:59 crc kubenswrapper[4998]: Trace[1694827292]: ---"Objects listed" error: 13770ms (16:03:59.731) Dec 03 16:03:59 crc kubenswrapper[4998]: Trace[1694827292]: [13.770394089s] [13.770394089s] END Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.731967 4998 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.732100 4998 trace.go:236] Trace[294307730]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 16:03:46.286) (total time: 13445ms): Dec 03 16:03:59 crc kubenswrapper[4998]: Trace[294307730]: ---"Objects listed" error: 13445ms (16:03:59.731) Dec 03 16:03:59 crc kubenswrapper[4998]: Trace[294307730]: [13.445959356s] [13.445959356s] END Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.732139 4998 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.733148 4998 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.733856 4998 trace.go:236] Trace[1863763716]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 16:03:45.402) (total time: 14331ms): Dec 03 16:03:59 crc kubenswrapper[4998]: Trace[1863763716]: ---"Objects listed" error: 14331ms (16:03:59.733) Dec 03 16:03:59 crc kubenswrapper[4998]: Trace[1863763716]: [14.331414277s] [14.331414277s] END Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.733883 4998 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.785620 4998 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": EOF" start-of-body= Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.785695 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": EOF" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.786176 4998 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection reset by peer" start-of-body= Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.786250 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection reset by peer" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.871892 4998 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.872247 4998 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.873740 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.873823 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.873842 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.873875 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.873893 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:03:59Z","lastTransitionTime":"2025-12-03T16:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:03:59 crc kubenswrapper[4998]: E1203 16:03:59.889869 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.895375 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.895417 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.895431 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.895448 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.895462 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:03:59Z","lastTransitionTime":"2025-12-03T16:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:03:59 crc kubenswrapper[4998]: E1203 16:03:59.909689 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.917134 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.917177 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.917191 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.917209 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.917226 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:03:59Z","lastTransitionTime":"2025-12-03T16:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:03:59 crc kubenswrapper[4998]: E1203 16:03:59.928601 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.931465 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.931505 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.931516 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.931534 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.931547 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:03:59Z","lastTransitionTime":"2025-12-03T16:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:03:59 crc kubenswrapper[4998]: E1203 16:03:59.940722 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.943333 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.943363 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.943372 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.943390 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.943400 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:03:59Z","lastTransitionTime":"2025-12-03T16:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:03:59 crc kubenswrapper[4998]: E1203 16:03:59.951391 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:03:59 crc kubenswrapper[4998]: E1203 16:03:59.951573 4998 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.953047 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.953072 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.953081 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.953096 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:03:59 crc kubenswrapper[4998]: I1203 16:03:59.953105 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:03:59Z","lastTransitionTime":"2025-12-03T16:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.054674 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.054712 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.054723 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.054739 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.054770 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:00Z","lastTransitionTime":"2025-12-03T16:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.156797 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.156824 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.156835 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.156848 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.156856 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:00Z","lastTransitionTime":"2025-12-03T16:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.259836 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.259875 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.259888 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.259904 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.259917 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:00Z","lastTransitionTime":"2025-12-03T16:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.362537 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.362585 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.362597 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.362617 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.362633 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:00Z","lastTransitionTime":"2025-12-03T16:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.465307 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.465380 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.465408 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.465440 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.465503 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:00Z","lastTransitionTime":"2025-12-03T16:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.569210 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.569301 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.569326 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.569358 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.569379 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:00Z","lastTransitionTime":"2025-12-03T16:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.607432 4998 apiserver.go:52] "Watching apiserver" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.611055 4998 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.616198 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-fcdxh","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-ovn-kubernetes/ovnkube-node-m2lh2","openshift-dns/node-resolver-9srmx","openshift-machine-config-operator/machine-config-daemon-9fjbr","openshift-multus/multus-additional-cni-plugins-qt6sw","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-operator/iptables-alerter-4ln5h","openshift-image-registry/node-ca-74tc7"] Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.617505 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.617916 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:00 crc kubenswrapper[4998]: E1203 16:04:00.618040 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.618135 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:00 crc kubenswrapper[4998]: E1203 16:04:00.618226 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.618320 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.618326 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.619048 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:00 crc kubenswrapper[4998]: E1203 16:04:00.619175 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.619446 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-9srmx" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.619657 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.619936 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.620089 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.620174 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-74tc7" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.620494 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.640677 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.640840 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.640897 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.641044 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.641133 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.641170 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.641245 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.641614 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.641914 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.641971 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.642129 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.642193 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.641934 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.642350 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.642570 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.642686 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.642811 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.642699 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.642922 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.642991 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.643026 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.643130 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.643179 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.643239 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.643465 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.645046 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.645353 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.645621 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.646082 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.646299 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.646886 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.647069 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.647211 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.647229 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.647301 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.671669 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.671727 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.671741 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.671781 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.671795 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:00Z","lastTransitionTime":"2025-12-03T16:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.702145 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.708392 4998 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.721606 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.734079 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.739939 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.739971 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.739991 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740014 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740030 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740047 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740067 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740083 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740103 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740119 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740138 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740237 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740259 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740275 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740290 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740306 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740322 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740343 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740359 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740375 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740426 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740442 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740463 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740479 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740496 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740483 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740512 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740530 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740517 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740547 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740621 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740652 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740675 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740703 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740743 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740759 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740788 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740841 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740862 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740883 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740902 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740919 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740952 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740973 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740977 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.740995 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.741014 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.742038 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.742040 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.742475 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.742544 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.742526 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.742689 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.743064 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.743143 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.743371 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.743458 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.743661 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.743809 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.743849 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.744011 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.744243 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.744266 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.744453 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: E1203 16:04:00.744595 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:04:01.241017627 +0000 UTC m=+19.852718040 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.744819 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.744852 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.744877 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.744900 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.744918 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.744937 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.744958 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.744979 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.744997 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.745018 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.745039 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.745057 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.745076 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.745097 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.745100 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.745115 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.745344 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.745370 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.745393 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.745416 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.745462 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.745532 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.745701 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.745790 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.745836 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.745884 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.745929 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.745973 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.746015 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.748252 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.749599 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.749715 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.749717 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.749808 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.749976 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750021 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750052 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750082 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750091 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750111 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750138 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750161 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750186 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750207 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750232 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750256 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750281 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750306 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750329 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750350 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750375 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750397 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750419 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750441 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750464 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750501 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750535 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750561 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750587 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750616 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750649 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750721 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750787 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750825 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750863 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750897 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750936 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.750978 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751018 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751045 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751068 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751092 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751136 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751159 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751224 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751305 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751492 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751536 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751561 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751586 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751613 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751636 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751661 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751693 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751726 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751772 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751800 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751827 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751852 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751876 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751898 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751923 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751946 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751971 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751997 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752021 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752045 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752128 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752160 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752184 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752213 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752254 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752293 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752329 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752364 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752403 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752435 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752469 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752502 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752551 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752591 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752628 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752657 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752689 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752724 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752807 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752846 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752901 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752930 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752979 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.753008 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.753058 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.753086 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.753136 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.753167 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.753234 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.753274 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.753332 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.753411 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.753470 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.753504 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.753557 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.753602 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.753683 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754316 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754365 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754392 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754418 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754446 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754472 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754499 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754531 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754557 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754583 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754608 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754633 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754658 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754688 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754723 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754757 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754800 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754826 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754853 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754882 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754981 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755011 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755037 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755111 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755196 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755239 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755264 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755445 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-systemd-units\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755495 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54d7l\" (UniqueName: \"kubernetes.io/projected/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-kube-api-access-54d7l\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755534 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-cni-netd\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755564 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-ovnkube-script-lib\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755610 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755647 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-cnibin\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755687 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755730 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xf9x\" (UniqueName: \"kubernetes.io/projected/25e154cc-ca57-431a-bce8-c9e245ec5a1e-kube-api-access-8xf9x\") pod \"multus-additional-cni-plugins-qt6sw\" (UID: \"25e154cc-ca57-431a-bce8-c9e245ec5a1e\") " pod="openshift-multus/multus-additional-cni-plugins-qt6sw" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755762 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-node-log\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755808 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/27aeb142-d4e3-4827-ac1f-c34de6822b14-rootfs\") pod \"machine-config-daemon-9fjbr\" (UID: \"27aeb142-d4e3-4827-ac1f-c34de6822b14\") " pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755833 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6htvc\" (UniqueName: \"kubernetes.io/projected/27aeb142-d4e3-4827-ac1f-c34de6822b14-kube-api-access-6htvc\") pod \"machine-config-daemon-9fjbr\" (UID: \"27aeb142-d4e3-4827-ac1f-c34de6822b14\") " pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755864 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b12292e6-f666-4bdc-8de8-334679d5f89a-host\") pod \"node-ca-74tc7\" (UID: \"b12292e6-f666-4bdc-8de8-334679d5f89a\") " pod="openshift-image-registry/node-ca-74tc7" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755890 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-run-ovn-kubernetes\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755922 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-host-run-k8s-cni-cncf-io\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755952 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755979 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-ovnkube-config\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756007 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-multus-conf-dir\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756505 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-etc-kubernetes\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756543 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qw5j\" (UniqueName: \"kubernetes.io/projected/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-kube-api-access-6qw5j\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756571 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/25e154cc-ca57-431a-bce8-c9e245ec5a1e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-qt6sw\" (UID: \"25e154cc-ca57-431a-bce8-c9e245ec5a1e\") " pod="openshift-multus/multus-additional-cni-plugins-qt6sw" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756601 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-cni-binary-copy\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756626 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-host-var-lib-kubelet\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756659 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756689 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/27aeb142-d4e3-4827-ac1f-c34de6822b14-mcd-auth-proxy-config\") pod \"machine-config-daemon-9fjbr\" (UID: \"27aeb142-d4e3-4827-ac1f-c34de6822b14\") " pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756727 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b12292e6-f666-4bdc-8de8-334679d5f89a-serviceca\") pod \"node-ca-74tc7\" (UID: \"b12292e6-f666-4bdc-8de8-334679d5f89a\") " pod="openshift-image-registry/node-ca-74tc7" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756782 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756825 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-hostroot\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756855 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw4tm\" (UniqueName: \"kubernetes.io/projected/5790060f-0fda-43a1-bde3-7332df05db85-kube-api-access-dw4tm\") pod \"node-resolver-9srmx\" (UID: \"5790060f-0fda-43a1-bde3-7332df05db85\") " pod="openshift-dns/node-resolver-9srmx" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756923 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756958 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756991 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/25e154cc-ca57-431a-bce8-c9e245ec5a1e-os-release\") pod \"multus-additional-cni-plugins-qt6sw\" (UID: \"25e154cc-ca57-431a-bce8-c9e245ec5a1e\") " pod="openshift-multus/multus-additional-cni-plugins-qt6sw" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757025 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-run-netns\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757055 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-var-lib-openvswitch\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757095 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757128 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-multus-cni-dir\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757162 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757189 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-os-release\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757216 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/25e154cc-ca57-431a-bce8-c9e245ec5a1e-system-cni-dir\") pod \"multus-additional-cni-plugins-qt6sw\" (UID: \"25e154cc-ca57-431a-bce8-c9e245ec5a1e\") " pod="openshift-multus/multus-additional-cni-plugins-qt6sw" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757246 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-kubelet\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757275 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-ovn-node-metrics-cert\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757301 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-run-ovn\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757339 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757369 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/5790060f-0fda-43a1-bde3-7332df05db85-hosts-file\") pod \"node-resolver-9srmx\" (UID: \"5790060f-0fda-43a1-bde3-7332df05db85\") " pod="openshift-dns/node-resolver-9srmx" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757393 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nftwj\" (UniqueName: \"kubernetes.io/projected/b12292e6-f666-4bdc-8de8-334679d5f89a-kube-api-access-nftwj\") pod \"node-ca-74tc7\" (UID: \"b12292e6-f666-4bdc-8de8-334679d5f89a\") " pod="openshift-image-registry/node-ca-74tc7" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757418 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-run-systemd\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757443 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-host-run-multus-certs\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757466 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/25e154cc-ca57-431a-bce8-c9e245ec5a1e-cnibin\") pod \"multus-additional-cni-plugins-qt6sw\" (UID: \"25e154cc-ca57-431a-bce8-c9e245ec5a1e\") " pod="openshift-multus/multus-additional-cni-plugins-qt6sw" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757490 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/25e154cc-ca57-431a-bce8-c9e245ec5a1e-cni-binary-copy\") pod \"multus-additional-cni-plugins-qt6sw\" (UID: \"25e154cc-ca57-431a-bce8-c9e245ec5a1e\") " pod="openshift-multus/multus-additional-cni-plugins-qt6sw" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757514 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757536 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-multus-socket-dir-parent\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757561 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757586 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757617 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-env-overrides\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757646 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757672 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/27aeb142-d4e3-4827-ac1f-c34de6822b14-proxy-tls\") pod \"machine-config-daemon-9fjbr\" (UID: \"27aeb142-d4e3-4827-ac1f-c34de6822b14\") " pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757703 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-etc-openvswitch\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757833 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-host-run-netns\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.751581 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752064 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752083 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757868 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-multus-daemon-config\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757921 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-log-socket\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758199 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/25e154cc-ca57-431a-bce8-c9e245ec5a1e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-qt6sw\" (UID: \"25e154cc-ca57-431a-bce8-c9e245ec5a1e\") " pod="openshift-multus/multus-additional-cni-plugins-qt6sw" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758261 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-slash\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758290 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-run-openvswitch\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758340 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-cni-bin\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758366 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-system-cni-dir\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758415 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-host-var-lib-cni-bin\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758447 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-host-var-lib-cni-multus\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758509 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758655 4998 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758720 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758738 4998 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758796 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758812 4998 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758828 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758870 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758888 4998 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758907 4998 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758951 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758968 4998 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758983 4998 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758999 4998 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.759039 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.759055 4998 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.759072 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.759124 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.759144 4998 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.759160 4998 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.759310 4998 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.759334 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.759406 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.759433 4998 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.759533 4998 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.759553 4998 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.759574 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.759600 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.759619 4998 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.759639 4998 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.759659 4998 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.759677 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.760851 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.760876 4998 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.762311 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.763243 4998 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.766851 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752094 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752202 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752231 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752498 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752519 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752572 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752621 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752836 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752873 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.752993 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.753126 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.753114 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.753189 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.753338 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.753363 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.753633 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.753640 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.753679 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.753720 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.753998 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754044 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754137 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754164 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754195 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754228 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.754274 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755622 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755513 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755716 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755841 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.755948 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756016 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756257 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756275 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756319 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756426 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756449 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.768026 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756474 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756588 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756567 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756616 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756540 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756754 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756783 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.756959 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757127 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757153 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757248 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757369 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757461 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757661 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757693 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757825 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757840 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.757914 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758016 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758263 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758282 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758540 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758549 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758565 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758640 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758829 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758873 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.758925 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.759355 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.759511 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.759569 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.768456 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.760232 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.760376 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.760452 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.760479 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.760549 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.761716 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.761883 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.762071 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.762122 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.762380 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.762596 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.762706 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.762778 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.762898 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.762948 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.762983 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.763069 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.763091 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.763183 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.763420 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.763458 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.763819 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.763925 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.763952 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.764156 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.764208 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.764492 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.764544 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.764563 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.764635 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.764823 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.764874 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.765000 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.765166 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.765207 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.765439 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.765519 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.765732 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.766027 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.766120 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.766178 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: E1203 16:04:00.766648 4998 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:04:00 crc kubenswrapper[4998]: E1203 16:04:00.769039 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:01.269019945 +0000 UTC m=+19.880720168 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:04:00 crc kubenswrapper[4998]: E1203 16:04:00.767515 4998 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:04:00 crc kubenswrapper[4998]: E1203 16:04:00.769196 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:01.269188339 +0000 UTC m=+19.880888562 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.775739 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.779537 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.779572 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.779581 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.779598 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.779607 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:00Z","lastTransitionTime":"2025-12-03T16:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:00 crc kubenswrapper[4998]: E1203 16:04:00.779714 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:04:00 crc kubenswrapper[4998]: E1203 16:04:00.779753 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:04:00 crc kubenswrapper[4998]: E1203 16:04:00.779789 4998 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.779797 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: E1203 16:04:00.779876 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:01.27985493 +0000 UTC m=+19.891555153 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:00 crc kubenswrapper[4998]: E1203 16:04:00.781464 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:04:00 crc kubenswrapper[4998]: E1203 16:04:00.781489 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:04:00 crc kubenswrapper[4998]: E1203 16:04:00.781501 4998 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.781597 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: E1203 16:04:00.781800 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:01.281783455 +0000 UTC m=+19.893483678 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.781800 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.782190 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.782251 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.785747 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.785956 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.788555 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.791048 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.791261 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.791337 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.791376 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.791826 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.791861 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.791931 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.792451 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.793464 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.794585 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.794773 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.794608 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.795133 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.795340 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.795515 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.795935 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.796114 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.796254 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.797031 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.797647 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.804277 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.805714 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.806871 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.806893 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.807505 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.807640 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.807746 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.807916 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.808274 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.808742 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.809814 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.810160 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.810330 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.810785 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.812126 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.815061 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.816010 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.818845 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.822460 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.822497 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.822804 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.823027 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.823526 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.823644 4998 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955" exitCode=255 Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.823702 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955"} Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.824267 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.824421 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.826524 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.827152 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.827255 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.830601 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.836572 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.841121 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.847773 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.848715 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.854193 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.858680 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.861235 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-systemd-units\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.861361 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-systemd-units\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.861798 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54d7l\" (UniqueName: \"kubernetes.io/projected/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-kube-api-access-54d7l\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.861858 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xf9x\" (UniqueName: \"kubernetes.io/projected/25e154cc-ca57-431a-bce8-c9e245ec5a1e-kube-api-access-8xf9x\") pod \"multus-additional-cni-plugins-qt6sw\" (UID: \"25e154cc-ca57-431a-bce8-c9e245ec5a1e\") " pod="openshift-multus/multus-additional-cni-plugins-qt6sw" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.861884 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-node-log\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.861903 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-cni-netd\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.861925 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-ovnkube-script-lib\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.861955 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-cnibin\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.861975 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-host-run-k8s-cni-cncf-io\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.861996 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/27aeb142-d4e3-4827-ac1f-c34de6822b14-rootfs\") pod \"machine-config-daemon-9fjbr\" (UID: \"27aeb142-d4e3-4827-ac1f-c34de6822b14\") " pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.862018 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6htvc\" (UniqueName: \"kubernetes.io/projected/27aeb142-d4e3-4827-ac1f-c34de6822b14-kube-api-access-6htvc\") pod \"machine-config-daemon-9fjbr\" (UID: \"27aeb142-d4e3-4827-ac1f-c34de6822b14\") " pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.862036 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b12292e6-f666-4bdc-8de8-334679d5f89a-host\") pod \"node-ca-74tc7\" (UID: \"b12292e6-f666-4bdc-8de8-334679d5f89a\") " pod="openshift-image-registry/node-ca-74tc7" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.862055 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-run-ovn-kubernetes\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.862070 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-node-log\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.862129 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-cni-netd\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.862107 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-host-run-k8s-cni-cncf-io\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.862200 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b12292e6-f666-4bdc-8de8-334679d5f89a-host\") pod \"node-ca-74tc7\" (UID: \"b12292e6-f666-4bdc-8de8-334679d5f89a\") " pod="openshift-image-registry/node-ca-74tc7" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.862260 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-cnibin\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.862294 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-run-ovn-kubernetes\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.862304 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/27aeb142-d4e3-4827-ac1f-c34de6822b14-rootfs\") pod \"machine-config-daemon-9fjbr\" (UID: \"27aeb142-d4e3-4827-ac1f-c34de6822b14\") " pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.862437 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-etc-kubernetes\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.862728 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-etc-kubernetes\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.862815 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qw5j\" (UniqueName: \"kubernetes.io/projected/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-kube-api-access-6qw5j\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.862865 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-ovnkube-config\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.862901 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-multus-conf-dir\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.862927 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/27aeb142-d4e3-4827-ac1f-c34de6822b14-mcd-auth-proxy-config\") pod \"machine-config-daemon-9fjbr\" (UID: \"27aeb142-d4e3-4827-ac1f-c34de6822b14\") " pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.862953 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/25e154cc-ca57-431a-bce8-c9e245ec5a1e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-qt6sw\" (UID: \"25e154cc-ca57-431a-bce8-c9e245ec5a1e\") " pod="openshift-multus/multus-additional-cni-plugins-qt6sw" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.862978 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-cni-binary-copy\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.863001 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-host-var-lib-kubelet\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.863025 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw4tm\" (UniqueName: \"kubernetes.io/projected/5790060f-0fda-43a1-bde3-7332df05db85-kube-api-access-dw4tm\") pod \"node-resolver-9srmx\" (UID: \"5790060f-0fda-43a1-bde3-7332df05db85\") " pod="openshift-dns/node-resolver-9srmx" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.863049 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b12292e6-f666-4bdc-8de8-334679d5f89a-serviceca\") pod \"node-ca-74tc7\" (UID: \"b12292e6-f666-4bdc-8de8-334679d5f89a\") " pod="openshift-image-registry/node-ca-74tc7" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.863085 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-hostroot\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.863111 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-multus-cni-dir\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.863118 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-ovnkube-script-lib\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.863137 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/25e154cc-ca57-431a-bce8-c9e245ec5a1e-os-release\") pod \"multus-additional-cni-plugins-qt6sw\" (UID: \"25e154cc-ca57-431a-bce8-c9e245ec5a1e\") " pod="openshift-multus/multus-additional-cni-plugins-qt6sw" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.863163 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-run-netns\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.863191 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-var-lib-openvswitch\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.863224 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-kubelet\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.863249 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-ovn-node-metrics-cert\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.863272 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-os-release\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.863332 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/25e154cc-ca57-431a-bce8-c9e245ec5a1e-system-cni-dir\") pod \"multus-additional-cni-plugins-qt6sw\" (UID: \"25e154cc-ca57-431a-bce8-c9e245ec5a1e\") " pod="openshift-multus/multus-additional-cni-plugins-qt6sw" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.863356 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/5790060f-0fda-43a1-bde3-7332df05db85-hosts-file\") pod \"node-resolver-9srmx\" (UID: \"5790060f-0fda-43a1-bde3-7332df05db85\") " pod="openshift-dns/node-resolver-9srmx" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.863377 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nftwj\" (UniqueName: \"kubernetes.io/projected/b12292e6-f666-4bdc-8de8-334679d5f89a-kube-api-access-nftwj\") pod \"node-ca-74tc7\" (UID: \"b12292e6-f666-4bdc-8de8-334679d5f89a\") " pod="openshift-image-registry/node-ca-74tc7" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.863429 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-run-ovn\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.863454 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/25e154cc-ca57-431a-bce8-c9e245ec5a1e-cnibin\") pod \"multus-additional-cni-plugins-qt6sw\" (UID: \"25e154cc-ca57-431a-bce8-c9e245ec5a1e\") " pod="openshift-multus/multus-additional-cni-plugins-qt6sw" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.864139 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-cni-binary-copy\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.864224 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-var-lib-openvswitch\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.864369 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-host-var-lib-kubelet\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.864940 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-multus-conf-dir\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.865765 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-ovnkube-config\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.866107 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-multus-cni-dir\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.866186 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-hostroot\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.866376 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/25e154cc-ca57-431a-bce8-c9e245ec5a1e-os-release\") pod \"multus-additional-cni-plugins-qt6sw\" (UID: \"25e154cc-ca57-431a-bce8-c9e245ec5a1e\") " pod="openshift-multus/multus-additional-cni-plugins-qt6sw" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.869566 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-os-release\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.869711 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-run-ovn\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.869757 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-kubelet\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.869857 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/5790060f-0fda-43a1-bde3-7332df05db85-hosts-file\") pod \"node-resolver-9srmx\" (UID: \"5790060f-0fda-43a1-bde3-7332df05db85\") " pod="openshift-dns/node-resolver-9srmx" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.869985 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/25e154cc-ca57-431a-bce8-c9e245ec5a1e-cnibin\") pod \"multus-additional-cni-plugins-qt6sw\" (UID: \"25e154cc-ca57-431a-bce8-c9e245ec5a1e\") " pod="openshift-multus/multus-additional-cni-plugins-qt6sw" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.870105 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/25e154cc-ca57-431a-bce8-c9e245ec5a1e-cni-binary-copy\") pod \"multus-additional-cni-plugins-qt6sw\" (UID: \"25e154cc-ca57-431a-bce8-c9e245ec5a1e\") " pod="openshift-multus/multus-additional-cni-plugins-qt6sw" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.870224 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-run-systemd\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.870326 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-host-run-multus-certs\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.870419 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-multus-socket-dir-parent\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.870519 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.870606 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.870702 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.870789 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-run-systemd\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.870831 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-host-run-multus-certs\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.870837 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.870789 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.870756 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-multus-socket-dir-parent\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.870718 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/27aeb142-d4e3-4827-ac1f-c34de6822b14-mcd-auth-proxy-config\") pod \"machine-config-daemon-9fjbr\" (UID: \"27aeb142-d4e3-4827-ac1f-c34de6822b14\") " pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.870973 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-run-netns\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.870981 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.870713 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/25e154cc-ca57-431a-bce8-c9e245ec5a1e-cni-binary-copy\") pod \"multus-additional-cni-plugins-qt6sw\" (UID: \"25e154cc-ca57-431a-bce8-c9e245ec5a1e\") " pod="openshift-multus/multus-additional-cni-plugins-qt6sw" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.872353 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-env-overrides\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.872462 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-host-run-netns\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.872562 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/27aeb142-d4e3-4827-ac1f-c34de6822b14-proxy-tls\") pod \"machine-config-daemon-9fjbr\" (UID: \"27aeb142-d4e3-4827-ac1f-c34de6822b14\") " pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.872644 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-host-run-netns\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.872727 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-etc-openvswitch\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.872844 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-multus-daemon-config\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.872922 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-etc-openvswitch\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.872597 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.873114 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-log-socket\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.872463 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/25e154cc-ca57-431a-bce8-c9e245ec5a1e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-qt6sw\" (UID: \"25e154cc-ca57-431a-bce8-c9e245ec5a1e\") " pod="openshift-multus/multus-additional-cni-plugins-qt6sw" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.872889 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-env-overrides\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.873249 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-log-socket\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.873377 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-host-var-lib-cni-multus\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.873473 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/25e154cc-ca57-431a-bce8-c9e245ec5a1e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-qt6sw\" (UID: \"25e154cc-ca57-431a-bce8-c9e245ec5a1e\") " pod="openshift-multus/multus-additional-cni-plugins-qt6sw" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.873560 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b12292e6-f666-4bdc-8de8-334679d5f89a-serviceca\") pod \"node-ca-74tc7\" (UID: \"b12292e6-f666-4bdc-8de8-334679d5f89a\") " pod="openshift-image-registry/node-ca-74tc7" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.873134 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/25e154cc-ca57-431a-bce8-c9e245ec5a1e-system-cni-dir\") pod \"multus-additional-cni-plugins-qt6sw\" (UID: \"25e154cc-ca57-431a-bce8-c9e245ec5a1e\") " pod="openshift-multus/multus-additional-cni-plugins-qt6sw" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.873632 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-host-var-lib-cni-multus\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.873570 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-slash\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.873844 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-run-openvswitch\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.873937 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-cni-bin\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.874037 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-system-cni-dir\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.874129 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-host-var-lib-cni-bin\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.874289 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-cni-bin\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.874091 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-multus-daemon-config\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.874505 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-slash\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.874508 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-host-var-lib-cni-bin\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.874678 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-run-openvswitch\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.874795 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-system-cni-dir\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.874923 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.875003 4998 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.875099 4998 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.875173 4998 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.875247 4998 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.875515 4998 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.875619 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.875715 4998 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.875811 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.875885 4998 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.875965 4998 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.876041 4998 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.876112 4998 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.876179 4998 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.876257 4998 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.876373 4998 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.876497 4998 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.876576 4998 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.877288 4998 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.877387 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.877468 4998 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.877586 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.877657 4998 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.877720 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.877820 4998 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.877906 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.878003 4998 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.878082 4998 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.878167 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.878248 4998 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.878326 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.878410 4998 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.878488 4998 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.878567 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.878647 4998 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.878723 4998 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.879034 4998 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.879129 4998 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.879209 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.879279 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.879359 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.880379 4998 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.880486 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.880566 4998 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.880643 4998 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.880725 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.880844 4998 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.881085 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.881179 4998 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.881236 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.881289 4998 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.881343 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.881396 4998 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.881456 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.881531 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.881641 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.881798 4998 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.881912 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.883061 4998 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.883190 4998 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.883292 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.883356 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.883411 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.883463 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.883521 4998 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.883582 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.883679 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.883739 4998 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.883820 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.883881 4998 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.883945 4998 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.884019 4998 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.884131 4998 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.884388 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.884470 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.884550 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.884618 4998 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.884690 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.884748 4998 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.884901 4998 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.884972 4998 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.885070 4998 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.885171 4998 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.885242 4998 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.885320 4998 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.885390 4998 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.885452 4998 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.885525 4998 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.885597 4998 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.885665 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.885722 4998 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.885808 4998 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.878941 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xf9x\" (UniqueName: \"kubernetes.io/projected/25e154cc-ca57-431a-bce8-c9e245ec5a1e-kube-api-access-8xf9x\") pod \"multus-additional-cni-plugins-qt6sw\" (UID: \"25e154cc-ca57-431a-bce8-c9e245ec5a1e\") " pod="openshift-multus/multus-additional-cni-plugins-qt6sw" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.878754 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qw5j\" (UniqueName: \"kubernetes.io/projected/bd9d66fb-a400-4810-aa7a-c81c9c24bd11-kube-api-access-6qw5j\") pod \"multus-fcdxh\" (UID: \"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\") " pod="openshift-multus/multus-fcdxh" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.876304 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/25e154cc-ca57-431a-bce8-c9e245ec5a1e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-qt6sw\" (UID: \"25e154cc-ca57-431a-bce8-c9e245ec5a1e\") " pod="openshift-multus/multus-additional-cni-plugins-qt6sw" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.884358 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.886005 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.880350 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw4tm\" (UniqueName: \"kubernetes.io/projected/5790060f-0fda-43a1-bde3-7332df05db85-kube-api-access-dw4tm\") pod \"node-resolver-9srmx\" (UID: \"5790060f-0fda-43a1-bde3-7332df05db85\") " pod="openshift-dns/node-resolver-9srmx" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.884542 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6htvc\" (UniqueName: \"kubernetes.io/projected/27aeb142-d4e3-4827-ac1f-c34de6822b14-kube-api-access-6htvc\") pod \"machine-config-daemon-9fjbr\" (UID: \"27aeb142-d4e3-4827-ac1f-c34de6822b14\") " pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.877559 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-ovn-node-metrics-cert\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.882655 4998 scope.go:117] "RemoveContainer" containerID="a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.883032 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.877352 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54d7l\" (UniqueName: \"kubernetes.io/projected/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-kube-api-access-54d7l\") pod \"ovnkube-node-m2lh2\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.881634 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.887915 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.888022 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.888041 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:00Z","lastTransitionTime":"2025-12-03T16:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.885912 4998 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.888156 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.888182 4998 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.888200 4998 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.888232 4998 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.888247 4998 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.888259 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.888280 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.888406 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nftwj\" (UniqueName: \"kubernetes.io/projected/b12292e6-f666-4bdc-8de8-334679d5f89a-kube-api-access-nftwj\") pod \"node-ca-74tc7\" (UID: \"b12292e6-f666-4bdc-8de8-334679d5f89a\") " pod="openshift-image-registry/node-ca-74tc7" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.878188 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/27aeb142-d4e3-4827-ac1f-c34de6822b14-proxy-tls\") pod \"machine-config-daemon-9fjbr\" (UID: \"27aeb142-d4e3-4827-ac1f-c34de6822b14\") " pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.888506 4998 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.888531 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.888545 4998 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.888558 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.888576 4998 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.888613 4998 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.888625 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.892994 4998 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893065 4998 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893078 4998 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893099 4998 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893129 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893139 4998 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893150 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893166 4998 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893177 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893210 4998 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893222 4998 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893237 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893246 4998 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893256 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893287 4998 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893298 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893308 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893317 4998 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893332 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893342 4998 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893369 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893378 4998 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893392 4998 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893410 4998 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893421 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893451 4998 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893466 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893477 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893486 4998 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893499 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893526 4998 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893536 4998 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893546 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893559 4998 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893568 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893577 4998 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893606 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893620 4998 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893630 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893640 4998 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893654 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893681 4998 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893693 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893702 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893715 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893726 4998 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893735 4998 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893779 4998 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893793 4998 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893801 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893810 4998 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893820 4998 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893832 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.893859 4998 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.909937 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.921134 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.929172 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.940503 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.947254 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.957964 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.960685 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.967447 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.975772 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:00 crc kubenswrapper[4998]: W1203 16:04:00.980117 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-c2b2187822b103dadda6835d6ae51ab9539ac8644043c4ab5b90c95c10fd30e1 WatchSource:0}: Error finding container c2b2187822b103dadda6835d6ae51ab9539ac8644043c4ab5b90c95c10fd30e1: Status 404 returned error can't find the container with id c2b2187822b103dadda6835d6ae51ab9539ac8644043c4ab5b90c95c10fd30e1 Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.981330 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.986174 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.994195 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.994247 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.994259 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.994278 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.994291 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:00Z","lastTransitionTime":"2025-12-03T16:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:00 crc kubenswrapper[4998]: I1203 16:04:00.995650 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:00.997003 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-9srmx" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.011170 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-fcdxh" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.015993 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.025217 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.031628 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.037484 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-74tc7" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.037839 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.043473 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.049718 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.063730 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.072810 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 16:04:01 crc kubenswrapper[4998]: W1203 16:04:01.078957 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27aeb142_d4e3_4827_ac1f_c34de6822b14.slice/crio-eb5e4aaa7bec49d7eb7895a9a19ba3eeefc428dc5ffff1b7ccccc5b78e2e8734 WatchSource:0}: Error finding container eb5e4aaa7bec49d7eb7895a9a19ba3eeefc428dc5ffff1b7ccccc5b78e2e8734: Status 404 returned error can't find the container with id eb5e4aaa7bec49d7eb7895a9a19ba3eeefc428dc5ffff1b7ccccc5b78e2e8734 Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.124257 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.124291 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.124303 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.124320 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.124330 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:01Z","lastTransitionTime":"2025-12-03T16:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.226511 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.226540 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.226549 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.226562 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.226573 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:01Z","lastTransitionTime":"2025-12-03T16:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.304139 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.304252 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.304276 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.304299 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.304321 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:01 crc kubenswrapper[4998]: E1203 16:04:01.304397 4998 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:04:01 crc kubenswrapper[4998]: E1203 16:04:01.304446 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:02.304430247 +0000 UTC m=+20.916130470 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:04:01 crc kubenswrapper[4998]: E1203 16:04:01.305652 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:04:01 crc kubenswrapper[4998]: E1203 16:04:01.305697 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:04:01 crc kubenswrapper[4998]: E1203 16:04:01.305712 4998 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:01 crc kubenswrapper[4998]: E1203 16:04:01.305798 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:02.305783489 +0000 UTC m=+20.917483712 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:01 crc kubenswrapper[4998]: E1203 16:04:01.305877 4998 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:04:01 crc kubenswrapper[4998]: E1203 16:04:01.305926 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:02.305909282 +0000 UTC m=+20.917609505 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:04:01 crc kubenswrapper[4998]: E1203 16:04:01.306004 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:04:02.305993694 +0000 UTC m=+20.917693917 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:04:01 crc kubenswrapper[4998]: E1203 16:04:01.306080 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:04:01 crc kubenswrapper[4998]: E1203 16:04:01.306091 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:04:01 crc kubenswrapper[4998]: E1203 16:04:01.306100 4998 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:01 crc kubenswrapper[4998]: E1203 16:04:01.306135 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:02.306119117 +0000 UTC m=+20.917819340 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.328838 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.328870 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.328878 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.328895 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.328904 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:01Z","lastTransitionTime":"2025-12-03T16:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.430888 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.431180 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.431189 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.431204 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.431214 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:01Z","lastTransitionTime":"2025-12-03T16:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.533655 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.533718 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.533733 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.533781 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.533797 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:01Z","lastTransitionTime":"2025-12-03T16:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.636659 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.636692 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.636704 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.636725 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.636738 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:01Z","lastTransitionTime":"2025-12-03T16:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.682063 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.682703 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.684235 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.685033 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.686283 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.687022 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.687748 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.689205 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.690022 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.691378 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.692067 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.693676 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.694406 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.695038 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.697684 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.698241 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.698846 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.699236 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.699869 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.700439 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.700938 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.701466 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.701901 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.702555 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.703024 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.703589 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.704192 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.704628 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.707405 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.707951 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.708717 4998 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.708825 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.710399 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.711300 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.711713 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.711709 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.713279 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.714226 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.714778 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.719042 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.719738 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.720322 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.721301 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.722398 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.723024 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.723833 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.724346 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.725700 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.726425 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.727316 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.727818 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.728265 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.729410 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.729976 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.730865 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.739352 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.739393 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.739405 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.739423 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.739435 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:01Z","lastTransitionTime":"2025-12-03T16:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.746128 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.791602 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.812289 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.828655 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-74tc7" event={"ID":"b12292e6-f666-4bdc-8de8-334679d5f89a","Type":"ContainerStarted","Data":"d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.828705 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-74tc7" event={"ID":"b12292e6-f666-4bdc-8de8-334679d5f89a","Type":"ContainerStarted","Data":"526ae043ffe380fdbe7c4ee52cc0dada8bccbee35fc30ca054e1ad77d9a57485"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.830417 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-9srmx" event={"ID":"5790060f-0fda-43a1-bde3-7332df05db85","Type":"ContainerStarted","Data":"dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.830443 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-9srmx" event={"ID":"5790060f-0fda-43a1-bde3-7332df05db85","Type":"ContainerStarted","Data":"f6d024d0b5c1132cbe3cc761a02b18ec50bff379c5dd03f41c13ae598dd68e8c"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.832047 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.833350 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.833854 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.835207 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.835231 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.835243 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"eb5e4aaa7bec49d7eb7895a9a19ba3eeefc428dc5ffff1b7ccccc5b78e2e8734"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.836797 4998 generic.go:334] "Generic (PLEG): container finished" podID="25e154cc-ca57-431a-bce8-c9e245ec5a1e" containerID="80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d" exitCode=0 Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.836843 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" event={"ID":"25e154cc-ca57-431a-bce8-c9e245ec5a1e","Type":"ContainerDied","Data":"80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.836858 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" event={"ID":"25e154cc-ca57-431a-bce8-c9e245ec5a1e","Type":"ContainerStarted","Data":"d6d842b219bcf76218e36087fa66d3ab00c27db7942c218b1a0638d168657c15"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.838536 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.838578 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"dd910ec96aa473c6212783c011b1413a26e20c863be317fbf0708a8e061d9889"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.839547 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"fc3f189bbbc5fa626826588c02030d9c974308d879da0ae5f2d8b23c9a94f83f"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.839981 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.843141 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.843166 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.843175 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.843190 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.843238 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:01Z","lastTransitionTime":"2025-12-03T16:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.843305 4998 generic.go:334] "Generic (PLEG): container finished" podID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerID="eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8" exitCode=0 Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.843388 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" event={"ID":"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9","Type":"ContainerDied","Data":"eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.843435 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" event={"ID":"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9","Type":"ContainerStarted","Data":"c740368518ed3aa02e3b48f5e955778758c5b3ba693eef7207af578dd11186ee"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.845059 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.845098 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.845113 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c2b2187822b103dadda6835d6ae51ab9539ac8644043c4ab5b90c95c10fd30e1"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.846166 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fcdxh" event={"ID":"bd9d66fb-a400-4810-aa7a-c81c9c24bd11","Type":"ContainerStarted","Data":"d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.846191 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fcdxh" event={"ID":"bd9d66fb-a400-4810-aa7a-c81c9c24bd11","Type":"ContainerStarted","Data":"5b10382101d46ff9f128a74be2707d9a70561127419d66552bbdb84fdf21e5fd"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.858115 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.874811 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.889073 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.900546 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.910175 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.920438 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.933339 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.945180 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.945221 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.945234 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.945254 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.945266 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:01Z","lastTransitionTime":"2025-12-03T16:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.947469 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.962500 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.974717 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.987602 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:01 crc kubenswrapper[4998]: I1203 16:04:01.999230 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.011467 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:02Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.023396 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:02Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.042956 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:02Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.047612 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.047649 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.047661 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.047679 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.047692 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:02Z","lastTransitionTime":"2025-12-03T16:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.055500 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:02Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.065820 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:02Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.080165 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:02Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.093481 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:02Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.105855 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:02Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.128116 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:02Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.158631 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.158673 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.158682 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.158700 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.158711 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:02Z","lastTransitionTime":"2025-12-03T16:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.262080 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.262120 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.262132 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.262149 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.262161 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:02Z","lastTransitionTime":"2025-12-03T16:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.314914 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.315015 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:02 crc kubenswrapper[4998]: E1203 16:04:02.315074 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:04:04.315036694 +0000 UTC m=+22.926736917 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:04:02 crc kubenswrapper[4998]: E1203 16:04:02.315124 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:04:02 crc kubenswrapper[4998]: E1203 16:04:02.315139 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:04:02 crc kubenswrapper[4998]: E1203 16:04:02.315150 4998 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:02 crc kubenswrapper[4998]: E1203 16:04:02.315197 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:04.315179958 +0000 UTC m=+22.926880181 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.315144 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:02 crc kubenswrapper[4998]: E1203 16:04:02.315257 4998 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.315277 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:02 crc kubenswrapper[4998]: E1203 16:04:02.315321 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:04.31529565 +0000 UTC m=+22.926995973 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.315350 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:02 crc kubenswrapper[4998]: E1203 16:04:02.315374 4998 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:04:02 crc kubenswrapper[4998]: E1203 16:04:02.315400 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:04.315392653 +0000 UTC m=+22.927092876 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:04:02 crc kubenswrapper[4998]: E1203 16:04:02.315476 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:04:02 crc kubenswrapper[4998]: E1203 16:04:02.315502 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:04:02 crc kubenswrapper[4998]: E1203 16:04:02.315515 4998 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:02 crc kubenswrapper[4998]: E1203 16:04:02.315558 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:04.315547626 +0000 UTC m=+22.927247959 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.364615 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.364657 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.364666 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.364683 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.364691 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:02Z","lastTransitionTime":"2025-12-03T16:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.468478 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.469000 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.469012 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.469033 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.469045 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:02Z","lastTransitionTime":"2025-12-03T16:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.571173 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.571213 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.571222 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.571238 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.571254 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:02Z","lastTransitionTime":"2025-12-03T16:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.673523 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.673567 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.673600 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.673646 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.673659 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:02Z","lastTransitionTime":"2025-12-03T16:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.676806 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.676850 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:02 crc kubenswrapper[4998]: E1203 16:04:02.676921 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.676806 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:02 crc kubenswrapper[4998]: E1203 16:04:02.676996 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:02 crc kubenswrapper[4998]: E1203 16:04:02.677061 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.775636 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.775674 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.775685 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.775701 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.775727 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:02Z","lastTransitionTime":"2025-12-03T16:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.854414 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" event={"ID":"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9","Type":"ContainerStarted","Data":"8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b"} Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.854472 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" event={"ID":"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9","Type":"ContainerStarted","Data":"91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423"} Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.854490 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" event={"ID":"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9","Type":"ContainerStarted","Data":"1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246"} Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.854502 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" event={"ID":"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9","Type":"ContainerStarted","Data":"ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335"} Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.854521 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" event={"ID":"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9","Type":"ContainerStarted","Data":"65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3"} Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.854535 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" event={"ID":"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9","Type":"ContainerStarted","Data":"aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894"} Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.856738 4998 generic.go:334] "Generic (PLEG): container finished" podID="25e154cc-ca57-431a-bce8-c9e245ec5a1e" containerID="8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c" exitCode=0 Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.856790 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" event={"ID":"25e154cc-ca57-431a-bce8-c9e245ec5a1e","Type":"ContainerDied","Data":"8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c"} Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.867894 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:02Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.878123 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.878172 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.878183 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.878203 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.878216 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:02Z","lastTransitionTime":"2025-12-03T16:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.878421 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:02Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.892067 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:02Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.902487 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:02Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.915610 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:02Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.932592 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:02Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.946860 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:02Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.959639 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:02Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.970853 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:02Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.981167 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.981365 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.981461 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.981609 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.981675 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:02Z","lastTransitionTime":"2025-12-03T16:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:02 crc kubenswrapper[4998]: I1203 16:04:02.989787 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:02Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.003536 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.017058 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.030524 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.084213 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.084256 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.084266 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.084281 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.084292 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:03Z","lastTransitionTime":"2025-12-03T16:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.093544 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.106976 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.108815 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.109114 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.119734 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.129519 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.141213 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.154378 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.168935 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.181198 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.186990 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.187026 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.187036 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.187079 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.187091 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:03Z","lastTransitionTime":"2025-12-03T16:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.192791 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.207084 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.221995 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.238616 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.252469 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.264302 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.277354 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.287401 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.291218 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.291334 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.291440 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.291498 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.291524 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:03Z","lastTransitionTime":"2025-12-03T16:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.301079 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.314365 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.328190 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.347199 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.360991 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.373057 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.386221 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.395003 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.395044 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.395057 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.395082 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.395097 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:03Z","lastTransitionTime":"2025-12-03T16:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.401722 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.421857 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.449161 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.463596 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.477628 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.497281 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.497325 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.497337 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.497356 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.497369 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:03Z","lastTransitionTime":"2025-12-03T16:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.600421 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.600463 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.600474 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.600489 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.600500 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:03Z","lastTransitionTime":"2025-12-03T16:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.702739 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.702844 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.702867 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.702893 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.702911 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:03Z","lastTransitionTime":"2025-12-03T16:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.806220 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.806281 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.806298 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.806321 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.806339 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:03Z","lastTransitionTime":"2025-12-03T16:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.862971 4998 generic.go:334] "Generic (PLEG): container finished" podID="25e154cc-ca57-431a-bce8-c9e245ec5a1e" containerID="301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97" exitCode=0 Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.863090 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" event={"ID":"25e154cc-ca57-431a-bce8-c9e245ec5a1e","Type":"ContainerDied","Data":"301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97"} Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.885006 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.905420 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.908862 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.908940 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.908963 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.908992 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.909015 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:03Z","lastTransitionTime":"2025-12-03T16:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.920091 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.934683 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.955281 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.978299 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:03 crc kubenswrapper[4998]: I1203 16:04:03.994457 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:03Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.008281 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.011081 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.011136 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.011154 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.011179 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.011198 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:04Z","lastTransitionTime":"2025-12-03T16:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.022333 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.035197 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.047170 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.062897 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.083667 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.114002 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.114060 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.114076 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.114100 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.114117 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:04Z","lastTransitionTime":"2025-12-03T16:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.123744 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.216475 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.216515 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.216530 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.216547 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.216558 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:04Z","lastTransitionTime":"2025-12-03T16:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.319278 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.319345 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.319363 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.319390 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.319411 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:04Z","lastTransitionTime":"2025-12-03T16:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.335596 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.335735 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.335802 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.335834 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.335860 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:04 crc kubenswrapper[4998]: E1203 16:04:04.335919 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:04:08.335886609 +0000 UTC m=+26.947586872 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:04:04 crc kubenswrapper[4998]: E1203 16:04:04.335943 4998 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:04:04 crc kubenswrapper[4998]: E1203 16:04:04.335971 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:04:04 crc kubenswrapper[4998]: E1203 16:04:04.336002 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:08.335984581 +0000 UTC m=+26.947684814 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:04:04 crc kubenswrapper[4998]: E1203 16:04:04.336009 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:04:04 crc kubenswrapper[4998]: E1203 16:04:04.336030 4998 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:04 crc kubenswrapper[4998]: E1203 16:04:04.336044 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:04:04 crc kubenswrapper[4998]: E1203 16:04:04.336069 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:04:04 crc kubenswrapper[4998]: E1203 16:04:04.336086 4998 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:04 crc kubenswrapper[4998]: E1203 16:04:04.336089 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:08.336070973 +0000 UTC m=+26.947771236 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:04 crc kubenswrapper[4998]: E1203 16:04:04.336133 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:08.336118665 +0000 UTC m=+26.947818918 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:04 crc kubenswrapper[4998]: E1203 16:04:04.336177 4998 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:04:04 crc kubenswrapper[4998]: E1203 16:04:04.336218 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:08.336205497 +0000 UTC m=+26.947905750 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.422439 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.422495 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.422512 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.422535 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.422553 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:04Z","lastTransitionTime":"2025-12-03T16:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.526114 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.526232 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.526299 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.526330 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.526395 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:04Z","lastTransitionTime":"2025-12-03T16:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.630428 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.630504 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.630521 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.630546 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.630565 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:04Z","lastTransitionTime":"2025-12-03T16:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.677405 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.677529 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.677564 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:04 crc kubenswrapper[4998]: E1203 16:04:04.677677 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:04 crc kubenswrapper[4998]: E1203 16:04:04.677900 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:04 crc kubenswrapper[4998]: E1203 16:04:04.678346 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.733326 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.733375 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.733392 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.733440 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.733458 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:04Z","lastTransitionTime":"2025-12-03T16:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.835627 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.835686 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.835702 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.835723 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.835740 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:04Z","lastTransitionTime":"2025-12-03T16:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.868857 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225"} Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.872897 4998 generic.go:334] "Generic (PLEG): container finished" podID="25e154cc-ca57-431a-bce8-c9e245ec5a1e" containerID="f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9" exitCode=0 Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.872972 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" event={"ID":"25e154cc-ca57-431a-bce8-c9e245ec5a1e","Type":"ContainerDied","Data":"f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9"} Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.890081 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.909739 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.926625 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.939726 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.940018 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.941211 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.941408 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.941552 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:04Z","lastTransitionTime":"2025-12-03T16:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.944675 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.971964 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:04 crc kubenswrapper[4998]: I1203 16:04:04.997632 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:04Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.017062 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.033317 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.044221 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.044298 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.044316 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.044338 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.044355 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:05Z","lastTransitionTime":"2025-12-03T16:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.050819 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.065065 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.080296 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.110387 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.135648 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.147962 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.147996 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.148007 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.148020 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.148030 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:05Z","lastTransitionTime":"2025-12-03T16:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.152198 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.187708 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.250699 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.250736 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.250745 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.250776 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.250786 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:05Z","lastTransitionTime":"2025-12-03T16:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.261848 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.273352 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.288221 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.301486 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.330523 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.351944 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.353457 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.353518 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.353535 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.353558 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.353576 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:05Z","lastTransitionTime":"2025-12-03T16:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.375063 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.389746 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.404731 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.424561 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.444755 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.455633 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.455683 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.455694 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.455713 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.455727 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:05Z","lastTransitionTime":"2025-12-03T16:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.465378 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.482685 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.558259 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.558313 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.558330 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.558354 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.558370 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:05Z","lastTransitionTime":"2025-12-03T16:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.661727 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.661836 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.661868 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.661894 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.661912 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:05Z","lastTransitionTime":"2025-12-03T16:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.764945 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.764987 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.764997 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.765010 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.765021 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:05Z","lastTransitionTime":"2025-12-03T16:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.867546 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.867606 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.867621 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.867639 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.867655 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:05Z","lastTransitionTime":"2025-12-03T16:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.878562 4998 generic.go:334] "Generic (PLEG): container finished" podID="25e154cc-ca57-431a-bce8-c9e245ec5a1e" containerID="7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678" exitCode=0 Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.878694 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" event={"ID":"25e154cc-ca57-431a-bce8-c9e245ec5a1e","Type":"ContainerDied","Data":"7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678"} Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.884271 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" event={"ID":"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9","Type":"ContainerStarted","Data":"3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538"} Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.900266 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.918860 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.934363 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.951673 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.969888 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.969941 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.969956 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.970014 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.970035 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:05Z","lastTransitionTime":"2025-12-03T16:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.976387 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:05 crc kubenswrapper[4998]: I1203 16:04:05.995065 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:05Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.010648 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.026901 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.041383 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.057746 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.070218 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.072014 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.072047 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.072058 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.072072 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.072081 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:06Z","lastTransitionTime":"2025-12-03T16:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.082470 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.095211 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.108925 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.162305 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.173506 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.175290 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.175345 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.175357 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.175377 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.175390 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:06Z","lastTransitionTime":"2025-12-03T16:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.175855 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.183220 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.200708 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.226175 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.244117 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.261218 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.277387 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.278503 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.278579 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.278603 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.278632 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.278657 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:06Z","lastTransitionTime":"2025-12-03T16:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.294112 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.316879 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.340842 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.354918 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.365702 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.378234 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.381142 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.381190 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.381207 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.381230 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.381248 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:06Z","lastTransitionTime":"2025-12-03T16:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.393099 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.409692 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.426634 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.442187 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.456013 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.469126 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.484614 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.484897 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.485060 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.485338 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.485599 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:06Z","lastTransitionTime":"2025-12-03T16:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.491938 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.512111 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d856098-4acb-4d57-8a27-de806955228c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53614cedb63f775d2b563e4405055cd6702c3f35980bc092ac07a3c9ff69aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c911081098fed822c164c7e59c1518ec11cf1c612e3ac4c84af8e694498c915d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9cdcf95d23503cce6d3fbc6de109783933aaaec4eefccd6f1834e234ed41f6e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.531732 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.549409 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.565729 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.578561 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.588546 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.588599 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.588615 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.588636 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.588649 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:06Z","lastTransitionTime":"2025-12-03T16:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.593363 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.608811 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.629654 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.666857 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.676764 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:06 crc kubenswrapper[4998]: E1203 16:04:06.676914 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.677073 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.677238 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:06 crc kubenswrapper[4998]: E1203 16:04:06.677279 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:06 crc kubenswrapper[4998]: E1203 16:04:06.677680 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.691463 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.691509 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.691519 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.691534 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.691544 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:06Z","lastTransitionTime":"2025-12-03T16:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.702152 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.794628 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.794900 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.794961 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.795069 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.795130 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:06Z","lastTransitionTime":"2025-12-03T16:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.893565 4998 generic.go:334] "Generic (PLEG): container finished" podID="25e154cc-ca57-431a-bce8-c9e245ec5a1e" containerID="2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300" exitCode=0 Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.893655 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" event={"ID":"25e154cc-ca57-431a-bce8-c9e245ec5a1e","Type":"ContainerDied","Data":"2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300"} Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.898014 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.898055 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.898072 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.898097 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.898116 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:06Z","lastTransitionTime":"2025-12-03T16:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.915029 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.934337 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.949580 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.966156 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:06 crc kubenswrapper[4998]: I1203 16:04:06.986663 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d856098-4acb-4d57-8a27-de806955228c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53614cedb63f775d2b563e4405055cd6702c3f35980bc092ac07a3c9ff69aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c911081098fed822c164c7e59c1518ec11cf1c612e3ac4c84af8e694498c915d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9cdcf95d23503cce6d3fbc6de109783933aaaec4eefccd6f1834e234ed41f6e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:06Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.006546 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.006602 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.006620 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.006647 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.006588 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.006664 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:07Z","lastTransitionTime":"2025-12-03T16:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.031343 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.048417 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.067795 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.107568 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.109537 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.109590 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.109603 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.109620 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.109633 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:07Z","lastTransitionTime":"2025-12-03T16:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.140936 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.183896 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.212586 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.212620 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.212632 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.212649 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.212662 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:07Z","lastTransitionTime":"2025-12-03T16:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.222296 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.263210 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.315403 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.315467 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.315486 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.315511 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.315530 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:07Z","lastTransitionTime":"2025-12-03T16:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.317372 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.418525 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.418594 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.418616 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.418644 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.418665 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:07Z","lastTransitionTime":"2025-12-03T16:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.522481 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.522531 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.522544 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.522560 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.522572 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:07Z","lastTransitionTime":"2025-12-03T16:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.624979 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.625055 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.625078 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.625107 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.625130 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:07Z","lastTransitionTime":"2025-12-03T16:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.727720 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.727781 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.727794 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.727809 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.727820 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:07Z","lastTransitionTime":"2025-12-03T16:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.830039 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.830096 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.830114 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.830149 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.830187 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:07Z","lastTransitionTime":"2025-12-03T16:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.903994 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" event={"ID":"25e154cc-ca57-431a-bce8-c9e245ec5a1e","Type":"ContainerStarted","Data":"50e48502ef54b1ba19755467c5f08d3596567a1f41f37ed5fdf117db7e617b7e"} Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.910197 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" event={"ID":"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9","Type":"ContainerStarted","Data":"1f172d9e0736e2c46658ab35b4626284cb594d4d028f83207ee91357c4104c80"} Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.910475 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.923057 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d856098-4acb-4d57-8a27-de806955228c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53614cedb63f775d2b563e4405055cd6702c3f35980bc092ac07a3c9ff69aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c911081098fed822c164c7e59c1518ec11cf1c612e3ac4c84af8e694498c915d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9cdcf95d23503cce6d3fbc6de109783933aaaec4eefccd6f1834e234ed41f6e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.932595 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.932668 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.932698 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.932728 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.932749 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:07Z","lastTransitionTime":"2025-12-03T16:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.938677 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.942413 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.962085 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e48502ef54b1ba19755467c5f08d3596567a1f41f37ed5fdf117db7e617b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.978680 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:07 crc kubenswrapper[4998]: I1203 16:04:07.997511 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:07Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.012462 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.026206 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.038282 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.038331 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.038348 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.038378 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.038396 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:08Z","lastTransitionTime":"2025-12-03T16:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.051742 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.070825 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.087353 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.102221 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.114118 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.122682 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.132713 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.141696 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.141737 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.141746 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.141777 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.141789 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:08Z","lastTransitionTime":"2025-12-03T16:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.145531 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.187599 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.208871 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.219711 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.229509 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.237899 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.243570 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.243607 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.243617 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.243631 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.243640 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:08Z","lastTransitionTime":"2025-12-03T16:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.259647 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f172d9e0736e2c46658ab35b4626284cb594d4d028f83207ee91357c4104c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.272507 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.282565 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.291577 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.300354 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.336782 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d856098-4acb-4d57-8a27-de806955228c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53614cedb63f775d2b563e4405055cd6702c3f35980bc092ac07a3c9ff69aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c911081098fed822c164c7e59c1518ec11cf1c612e3ac4c84af8e694498c915d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9cdcf95d23503cce6d3fbc6de109783933aaaec4eefccd6f1834e234ed41f6e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.345190 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.345235 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.345247 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.345266 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.345299 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:08Z","lastTransitionTime":"2025-12-03T16:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.372811 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.372929 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.372958 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:08 crc kubenswrapper[4998]: E1203 16:04:08.373030 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:04:16.372998802 +0000 UTC m=+34.984699065 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:04:08 crc kubenswrapper[4998]: E1203 16:04:08.373050 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:04:08 crc kubenswrapper[4998]: E1203 16:04:08.373067 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:04:08 crc kubenswrapper[4998]: E1203 16:04:08.373077 4998 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:08 crc kubenswrapper[4998]: E1203 16:04:08.373119 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:16.373106905 +0000 UTC m=+34.984807128 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.373115 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.373160 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:08 crc kubenswrapper[4998]: E1203 16:04:08.373189 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:04:08 crc kubenswrapper[4998]: E1203 16:04:08.373254 4998 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:04:08 crc kubenswrapper[4998]: E1203 16:04:08.373264 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:04:08 crc kubenswrapper[4998]: E1203 16:04:08.373290 4998 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:08 crc kubenswrapper[4998]: E1203 16:04:08.373305 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:16.373290539 +0000 UTC m=+34.984990792 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:04:08 crc kubenswrapper[4998]: E1203 16:04:08.373328 4998 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:04:08 crc kubenswrapper[4998]: E1203 16:04:08.373394 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:16.373363771 +0000 UTC m=+34.985064034 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:08 crc kubenswrapper[4998]: E1203 16:04:08.373432 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:16.373414702 +0000 UTC m=+34.985115025 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.382099 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.420347 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e48502ef54b1ba19755467c5f08d3596567a1f41f37ed5fdf117db7e617b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.447917 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.447983 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.447996 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.448018 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.448032 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:08Z","lastTransitionTime":"2025-12-03T16:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.461318 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.504930 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.550107 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.550148 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.550157 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.550171 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.550180 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:08Z","lastTransitionTime":"2025-12-03T16:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.653908 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.653987 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.654003 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.654027 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.654049 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:08Z","lastTransitionTime":"2025-12-03T16:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.677375 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.677423 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.677424 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:08 crc kubenswrapper[4998]: E1203 16:04:08.677601 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:08 crc kubenswrapper[4998]: E1203 16:04:08.677712 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:08 crc kubenswrapper[4998]: E1203 16:04:08.677869 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.757250 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.757295 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.757307 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.757325 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.757337 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:08Z","lastTransitionTime":"2025-12-03T16:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.860251 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.860316 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.860343 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.860371 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.860392 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:08Z","lastTransitionTime":"2025-12-03T16:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.914063 4998 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.914611 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.943932 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.958746 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d856098-4acb-4d57-8a27-de806955228c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53614cedb63f775d2b563e4405055cd6702c3f35980bc092ac07a3c9ff69aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c911081098fed822c164c7e59c1518ec11cf1c612e3ac4c84af8e694498c915d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9cdcf95d23503cce6d3fbc6de109783933aaaec4eefccd6f1834e234ed41f6e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.962239 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.962285 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.962295 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.962312 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.962323 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:08Z","lastTransitionTime":"2025-12-03T16:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.976177 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:08 crc kubenswrapper[4998]: I1203 16:04:08.997642 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e48502ef54b1ba19755467c5f08d3596567a1f41f37ed5fdf117db7e617b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:08Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.015553 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.029788 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.059670 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.065174 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.065241 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.065273 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.065305 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.065329 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:09Z","lastTransitionTime":"2025-12-03T16:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.087115 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.107867 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.130170 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.148548 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.168343 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.168391 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.168403 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.168420 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.168432 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:09Z","lastTransitionTime":"2025-12-03T16:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.174311 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f172d9e0736e2c46658ab35b4626284cb594d4d028f83207ee91357c4104c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.199278 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.213587 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.236329 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.255254 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:09Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.270742 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.270786 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.270795 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.270808 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.270817 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:09Z","lastTransitionTime":"2025-12-03T16:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.373254 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.373298 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.373312 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.373328 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.373338 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:09Z","lastTransitionTime":"2025-12-03T16:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.476082 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.476127 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.476143 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.476163 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.476178 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:09Z","lastTransitionTime":"2025-12-03T16:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.578671 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.578722 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.578733 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.578767 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.578779 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:09Z","lastTransitionTime":"2025-12-03T16:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.680964 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.681024 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.681041 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.681068 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.681086 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:09Z","lastTransitionTime":"2025-12-03T16:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.784285 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.784348 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.784367 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.784392 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.784410 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:09Z","lastTransitionTime":"2025-12-03T16:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.887307 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.887435 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.887468 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.887505 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.887531 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:09Z","lastTransitionTime":"2025-12-03T16:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.917624 4998 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.989428 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.989474 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.989485 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.989502 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:09 crc kubenswrapper[4998]: I1203 16:04:09.989513 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:09Z","lastTransitionTime":"2025-12-03T16:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.092389 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.092436 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.092448 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.092491 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.092505 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:10Z","lastTransitionTime":"2025-12-03T16:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.137530 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.137579 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.137597 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.137623 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.137642 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:10Z","lastTransitionTime":"2025-12-03T16:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:10 crc kubenswrapper[4998]: E1203 16:04:10.157532 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.163018 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.163064 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.163075 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.163094 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.163109 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:10Z","lastTransitionTime":"2025-12-03T16:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:10 crc kubenswrapper[4998]: E1203 16:04:10.184698 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.188495 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.188527 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.188535 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.188550 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.188561 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:10Z","lastTransitionTime":"2025-12-03T16:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:10 crc kubenswrapper[4998]: E1203 16:04:10.199980 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.204632 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.204698 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.204710 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.204730 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.204792 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:10Z","lastTransitionTime":"2025-12-03T16:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:10 crc kubenswrapper[4998]: E1203 16:04:10.224877 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.229628 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.229695 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.229721 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.229748 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.229787 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:10Z","lastTransitionTime":"2025-12-03T16:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:10 crc kubenswrapper[4998]: E1203 16:04:10.250698 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:10 crc kubenswrapper[4998]: E1203 16:04:10.250970 4998 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.252636 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.252697 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.252727 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.252784 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.252802 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:10Z","lastTransitionTime":"2025-12-03T16:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.356325 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.356392 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.356404 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.356422 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.356434 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:10Z","lastTransitionTime":"2025-12-03T16:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.459247 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.459346 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.459365 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.459393 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.459412 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:10Z","lastTransitionTime":"2025-12-03T16:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.562831 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.562896 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.562908 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.562925 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.562960 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:10Z","lastTransitionTime":"2025-12-03T16:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.665042 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.665113 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.665139 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.665170 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.665194 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:10Z","lastTransitionTime":"2025-12-03T16:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.677408 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:10 crc kubenswrapper[4998]: E1203 16:04:10.677551 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.677590 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:10 crc kubenswrapper[4998]: E1203 16:04:10.677710 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.678022 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:10 crc kubenswrapper[4998]: E1203 16:04:10.678368 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.768466 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.768533 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.768555 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.768578 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.768593 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:10Z","lastTransitionTime":"2025-12-03T16:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.871810 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.871894 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.871925 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.871955 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.871972 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:10Z","lastTransitionTime":"2025-12-03T16:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.924291 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m2lh2_4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9/ovnkube-controller/0.log" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.928022 4998 generic.go:334] "Generic (PLEG): container finished" podID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerID="1f172d9e0736e2c46658ab35b4626284cb594d4d028f83207ee91357c4104c80" exitCode=1 Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.928080 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" event={"ID":"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9","Type":"ContainerDied","Data":"1f172d9e0736e2c46658ab35b4626284cb594d4d028f83207ee91357c4104c80"} Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.928731 4998 scope.go:117] "RemoveContainer" containerID="1f172d9e0736e2c46658ab35b4626284cb594d4d028f83207ee91357c4104c80" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.948111 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.971932 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.973849 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.973901 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.973917 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.973939 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.973959 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:10Z","lastTransitionTime":"2025-12-03T16:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:10 crc kubenswrapper[4998]: I1203 16:04:10.986526 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.000910 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:10Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.016657 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.043825 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f172d9e0736e2c46658ab35b4626284cb594d4d028f83207ee91357c4104c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f172d9e0736e2c46658ab35b4626284cb594d4d028f83207ee91357c4104c80\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"message\\\":\\\"er 1 for removal\\\\nI1203 16:04:10.277403 6334 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 16:04:10.277454 6334 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 16:04:10.277498 6334 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 16:04:10.277518 6334 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:10.277599 6334 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 16:04:10.277626 6334 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:04:10.277630 6334 factory.go:656] Stopping watch factory\\\\nI1203 16:04:10.277639 6334 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:04:10.277658 6334 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 16:04:10.277677 6334 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16:04:10.277701 6334 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:04:10.276934 6334 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:10.277716 6334 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.075848 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.077010 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.077046 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.077057 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.077075 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.077088 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:11Z","lastTransitionTime":"2025-12-03T16:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.100142 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.120116 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.139287 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.152451 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.166326 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.179075 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.179119 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.179131 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.179150 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.179168 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:11Z","lastTransitionTime":"2025-12-03T16:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.188114 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e48502ef54b1ba19755467c5f08d3596567a1f41f37ed5fdf117db7e617b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.204711 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d856098-4acb-4d57-8a27-de806955228c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53614cedb63f775d2b563e4405055cd6702c3f35980bc092ac07a3c9ff69aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c911081098fed822c164c7e59c1518ec11cf1c612e3ac4c84af8e694498c915d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9cdcf95d23503cce6d3fbc6de109783933aaaec4eefccd6f1834e234ed41f6e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.221168 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.281959 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.281997 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.282006 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.282021 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.282030 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:11Z","lastTransitionTime":"2025-12-03T16:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.384723 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.384788 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.384800 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.384822 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.384848 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:11Z","lastTransitionTime":"2025-12-03T16:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.488309 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.488363 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.488379 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.488400 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.488414 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:11Z","lastTransitionTime":"2025-12-03T16:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.590687 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.590720 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.590728 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.590743 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.590769 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:11Z","lastTransitionTime":"2025-12-03T16:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.693289 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.693405 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.693447 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.693465 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.693489 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.693505 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:11Z","lastTransitionTime":"2025-12-03T16:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.714132 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e48502ef54b1ba19755467c5f08d3596567a1f41f37ed5fdf117db7e617b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.727084 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d856098-4acb-4d57-8a27-de806955228c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53614cedb63f775d2b563e4405055cd6702c3f35980bc092ac07a3c9ff69aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c911081098fed822c164c7e59c1518ec11cf1c612e3ac4c84af8e694498c915d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9cdcf95d23503cce6d3fbc6de109783933aaaec4eefccd6f1834e234ed41f6e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.745975 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.763806 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.780233 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.795817 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.795867 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.795880 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.795901 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.795913 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:11Z","lastTransitionTime":"2025-12-03T16:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.801612 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.819429 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.834069 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.856846 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f172d9e0736e2c46658ab35b4626284cb594d4d028f83207ee91357c4104c80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f172d9e0736e2c46658ab35b4626284cb594d4d028f83207ee91357c4104c80\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"message\\\":\\\"er 1 for removal\\\\nI1203 16:04:10.277403 6334 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 16:04:10.277454 6334 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 16:04:10.277498 6334 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 16:04:10.277518 6334 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:10.277599 6334 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 16:04:10.277626 6334 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:04:10.277630 6334 factory.go:656] Stopping watch factory\\\\nI1203 16:04:10.277639 6334 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:04:10.277658 6334 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 16:04:10.277677 6334 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16:04:10.277701 6334 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:04:10.276934 6334 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:10.277716 6334 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.881130 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.898686 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.898937 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.898983 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.899003 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.899026 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.899044 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:11Z","lastTransitionTime":"2025-12-03T16:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.918938 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.935719 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m2lh2_4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9/ovnkube-controller/0.log" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.937384 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.939914 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" event={"ID":"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9","Type":"ContainerStarted","Data":"609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95"} Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.940091 4998 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.952061 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.972951 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d856098-4acb-4d57-8a27-de806955228c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53614cedb63f775d2b563e4405055cd6702c3f35980bc092ac07a3c9ff69aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c911081098fed822c164c7e59c1518ec11cf1c612e3ac4c84af8e694498c915d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9cdcf95d23503cce6d3fbc6de109783933aaaec4eefccd6f1834e234ed41f6e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:11 crc kubenswrapper[4998]: I1203 16:04:11.991230 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:11Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.002627 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.002695 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.002712 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.002984 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.003003 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:12Z","lastTransitionTime":"2025-12-03T16:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.016503 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e48502ef54b1ba19755467c5f08d3596567a1f41f37ed5fdf117db7e617b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.033920 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.054125 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.090905 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.106309 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.106353 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.106364 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.106381 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.106394 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:12Z","lastTransitionTime":"2025-12-03T16:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.113272 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.132606 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.152648 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.166319 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.195690 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f172d9e0736e2c46658ab35b4626284cb594d4d028f83207ee91357c4104c80\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"message\\\":\\\"er 1 for removal\\\\nI1203 16:04:10.277403 6334 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 16:04:10.277454 6334 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 16:04:10.277498 6334 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 16:04:10.277518 6334 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:10.277599 6334 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 16:04:10.277626 6334 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:04:10.277630 6334 factory.go:656] Stopping watch factory\\\\nI1203 16:04:10.277639 6334 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:04:10.277658 6334 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 16:04:10.277677 6334 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16:04:10.277701 6334 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:04:10.276934 6334 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:10.277716 6334 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.209960 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.210014 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.210030 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.210053 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.210072 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:12Z","lastTransitionTime":"2025-12-03T16:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.213629 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.232020 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.244695 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.260451 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.313133 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.313191 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.313206 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.313227 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.313243 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:12Z","lastTransitionTime":"2025-12-03T16:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.416114 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.416532 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.416664 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.416874 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.417037 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:12Z","lastTransitionTime":"2025-12-03T16:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.520279 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.520315 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.520323 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.520338 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.520348 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:12Z","lastTransitionTime":"2025-12-03T16:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.622382 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.623045 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.623096 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.623125 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.623144 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:12Z","lastTransitionTime":"2025-12-03T16:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.677083 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.677203 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:12 crc kubenswrapper[4998]: E1203 16:04:12.677276 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:12 crc kubenswrapper[4998]: E1203 16:04:12.677349 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.677207 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:12 crc kubenswrapper[4998]: E1203 16:04:12.677564 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.726201 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.726279 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.726304 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.726339 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.726366 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:12Z","lastTransitionTime":"2025-12-03T16:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.746295 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d"] Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.746722 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.748787 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.750396 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.762881 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.787991 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.806562 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.818157 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj529\" (UniqueName: \"kubernetes.io/projected/b65cf533-fd60-47a3-aab2-b7377bab9f9d-kube-api-access-zj529\") pod \"ovnkube-control-plane-749d76644c-47k9d\" (UID: \"b65cf533-fd60-47a3-aab2-b7377bab9f9d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.818251 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b65cf533-fd60-47a3-aab2-b7377bab9f9d-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-47k9d\" (UID: \"b65cf533-fd60-47a3-aab2-b7377bab9f9d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.818306 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b65cf533-fd60-47a3-aab2-b7377bab9f9d-env-overrides\") pod \"ovnkube-control-plane-749d76644c-47k9d\" (UID: \"b65cf533-fd60-47a3-aab2-b7377bab9f9d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.818445 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b65cf533-fd60-47a3-aab2-b7377bab9f9d-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-47k9d\" (UID: \"b65cf533-fd60-47a3-aab2-b7377bab9f9d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.826497 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.828309 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.828335 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.828343 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.828356 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.828364 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:12Z","lastTransitionTime":"2025-12-03T16:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.853266 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f172d9e0736e2c46658ab35b4626284cb594d4d028f83207ee91357c4104c80\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"message\\\":\\\"er 1 for removal\\\\nI1203 16:04:10.277403 6334 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 16:04:10.277454 6334 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 16:04:10.277498 6334 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 16:04:10.277518 6334 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:10.277599 6334 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 16:04:10.277626 6334 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:04:10.277630 6334 factory.go:656] Stopping watch factory\\\\nI1203 16:04:10.277639 6334 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:04:10.277658 6334 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 16:04:10.277677 6334 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16:04:10.277701 6334 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:04:10.276934 6334 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:10.277716 6334 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.884513 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.903593 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.915649 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.919920 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b65cf533-fd60-47a3-aab2-b7377bab9f9d-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-47k9d\" (UID: \"b65cf533-fd60-47a3-aab2-b7377bab9f9d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.919975 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj529\" (UniqueName: \"kubernetes.io/projected/b65cf533-fd60-47a3-aab2-b7377bab9f9d-kube-api-access-zj529\") pod \"ovnkube-control-plane-749d76644c-47k9d\" (UID: \"b65cf533-fd60-47a3-aab2-b7377bab9f9d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.919998 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b65cf533-fd60-47a3-aab2-b7377bab9f9d-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-47k9d\" (UID: \"b65cf533-fd60-47a3-aab2-b7377bab9f9d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.920018 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b65cf533-fd60-47a3-aab2-b7377bab9f9d-env-overrides\") pod \"ovnkube-control-plane-749d76644c-47k9d\" (UID: \"b65cf533-fd60-47a3-aab2-b7377bab9f9d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.920847 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b65cf533-fd60-47a3-aab2-b7377bab9f9d-env-overrides\") pod \"ovnkube-control-plane-749d76644c-47k9d\" (UID: \"b65cf533-fd60-47a3-aab2-b7377bab9f9d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.920934 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b65cf533-fd60-47a3-aab2-b7377bab9f9d-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-47k9d\" (UID: \"b65cf533-fd60-47a3-aab2-b7377bab9f9d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.927413 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b65cf533-fd60-47a3-aab2-b7377bab9f9d-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-47k9d\" (UID: \"b65cf533-fd60-47a3-aab2-b7377bab9f9d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.928830 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.930480 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.930509 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.930521 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.930537 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.930549 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:12Z","lastTransitionTime":"2025-12-03T16:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.938836 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj529\" (UniqueName: \"kubernetes.io/projected/b65cf533-fd60-47a3-aab2-b7377bab9f9d-kube-api-access-zj529\") pod \"ovnkube-control-plane-749d76644c-47k9d\" (UID: \"b65cf533-fd60-47a3-aab2-b7377bab9f9d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.944189 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m2lh2_4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9/ovnkube-controller/1.log" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.944742 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.944980 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m2lh2_4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9/ovnkube-controller/0.log" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.947493 4998 generic.go:334] "Generic (PLEG): container finished" podID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerID="609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95" exitCode=1 Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.947533 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" event={"ID":"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9","Type":"ContainerDied","Data":"609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95"} Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.947571 4998 scope.go:117] "RemoveContainer" containerID="1f172d9e0736e2c46658ab35b4626284cb594d4d028f83207ee91357c4104c80" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.948296 4998 scope.go:117] "RemoveContainer" containerID="609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95" Dec 03 16:04:12 crc kubenswrapper[4998]: E1203 16:04:12.948454 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-m2lh2_openshift-ovn-kubernetes(4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.960513 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.978380 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:12 crc kubenswrapper[4998]: I1203 16:04:12.993944 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65cf533-fd60-47a3-aab2-b7377bab9f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47k9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:12Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.014550 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d856098-4acb-4d57-8a27-de806955228c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53614cedb63f775d2b563e4405055cd6702c3f35980bc092ac07a3c9ff69aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c911081098fed822c164c7e59c1518ec11cf1c612e3ac4c84af8e694498c915d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9cdcf95d23503cce6d3fbc6de109783933aaaec4eefccd6f1834e234ed41f6e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.025973 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.033238 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.033294 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.033306 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.033323 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.033337 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:13Z","lastTransitionTime":"2025-12-03T16:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.041382 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e48502ef54b1ba19755467c5f08d3596567a1f41f37ed5fdf117db7e617b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.050230 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.058095 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.068868 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.070832 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.084273 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.099332 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d856098-4acb-4d57-8a27-de806955228c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53614cedb63f775d2b563e4405055cd6702c3f35980bc092ac07a3c9ff69aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c911081098fed822c164c7e59c1518ec11cf1c612e3ac4c84af8e694498c915d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9cdcf95d23503cce6d3fbc6de109783933aaaec4eefccd6f1834e234ed41f6e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.116617 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.133224 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e48502ef54b1ba19755467c5f08d3596567a1f41f37ed5fdf117db7e617b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.136225 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.136248 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.136256 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.136269 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.136278 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:13Z","lastTransitionTime":"2025-12-03T16:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.150436 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65cf533-fd60-47a3-aab2-b7377bab9f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47k9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.168005 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.182136 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.194378 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.214493 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f172d9e0736e2c46658ab35b4626284cb594d4d028f83207ee91357c4104c80\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:04:10Z\\\",\\\"message\\\":\\\"er 1 for removal\\\\nI1203 16:04:10.277403 6334 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 16:04:10.277454 6334 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 16:04:10.277498 6334 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 16:04:10.277518 6334 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:10.277599 6334 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 16:04:10.277626 6334 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:04:10.277630 6334 factory.go:656] Stopping watch factory\\\\nI1203 16:04:10.277639 6334 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:04:10.277658 6334 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 16:04:10.277677 6334 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16:04:10.277701 6334 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:04:10.276934 6334 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:10.277716 6334 handler.go:208] Removed *v1.NetworkPolicy ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"message\\\":\\\"ler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 16:04:11.823033 6464 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:04:11.823064 6464 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:04:11.823080 6464 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:04:11.823178 6464 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:11.823309 6464 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823440 6464 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823488 6464 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823628 6464 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:04:11.823784 6464 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:11.824254 6464 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.232418 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.239435 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.239512 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.239524 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.239537 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.239546 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:13Z","lastTransitionTime":"2025-12-03T16:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.246112 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.258872 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.270571 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.343265 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.343298 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.343342 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.343359 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.343370 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:13Z","lastTransitionTime":"2025-12-03T16:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.445477 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.445503 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.445513 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.445530 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.445543 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:13Z","lastTransitionTime":"2025-12-03T16:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.548375 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.548584 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.548662 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.548766 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.548838 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:13Z","lastTransitionTime":"2025-12-03T16:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.651140 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.651201 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.651219 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.651245 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.651274 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:13Z","lastTransitionTime":"2025-12-03T16:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.754311 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.754693 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.754916 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.755129 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.755325 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:13Z","lastTransitionTime":"2025-12-03T16:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.858691 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.858785 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.858804 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.858832 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.858850 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:13Z","lastTransitionTime":"2025-12-03T16:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.954453 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m2lh2_4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9/ovnkube-controller/1.log" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.962037 4998 scope.go:117] "RemoveContainer" containerID="609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95" Dec 03 16:04:13 crc kubenswrapper[4998]: E1203 16:04:13.962742 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-m2lh2_openshift-ovn-kubernetes(4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.963865 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.963924 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.963952 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.963998 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.964024 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:13Z","lastTransitionTime":"2025-12-03T16:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.967614 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" event={"ID":"b65cf533-fd60-47a3-aab2-b7377bab9f9d","Type":"ContainerStarted","Data":"ec93a183a13918fa4a6c42b6d0feed52f46b3eac00d85977936772b4cd7e4cb9"} Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.967665 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" event={"ID":"b65cf533-fd60-47a3-aab2-b7377bab9f9d","Type":"ContainerStarted","Data":"44d20df9678d9388f9611bf182cf962781bbcb668bccef2fba40783ca6cd7cbd"} Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.967675 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" event={"ID":"b65cf533-fd60-47a3-aab2-b7377bab9f9d","Type":"ContainerStarted","Data":"de8363eda95866036930710ef34edb4703263a829a557a9309c05505a52614d4"} Dec 03 16:04:13 crc kubenswrapper[4998]: I1203 16:04:13.984286 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:13Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.004915 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.036989 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"message\\\":\\\"ler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 16:04:11.823033 6464 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:04:11.823064 6464 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:04:11.823080 6464 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:04:11.823178 6464 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:11.823309 6464 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823440 6464 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823488 6464 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823628 6464 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:04:11.823784 6464 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:11.824254 6464 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-m2lh2_openshift-ovn-kubernetes(4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.067323 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.067404 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.067425 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.067451 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.067468 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:14Z","lastTransitionTime":"2025-12-03T16:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.071485 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.095614 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.118116 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.138790 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.158143 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.170126 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.170169 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.170185 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.170211 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.170230 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:14Z","lastTransitionTime":"2025-12-03T16:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.174807 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.193508 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.211725 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.225963 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.243161 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d856098-4acb-4d57-8a27-de806955228c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53614cedb63f775d2b563e4405055cd6702c3f35980bc092ac07a3c9ff69aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c911081098fed822c164c7e59c1518ec11cf1c612e3ac4c84af8e694498c915d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9cdcf95d23503cce6d3fbc6de109783933aaaec4eefccd6f1834e234ed41f6e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.261120 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.273698 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.273782 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.273804 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.273830 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.273851 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:14Z","lastTransitionTime":"2025-12-03T16:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.281650 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e48502ef54b1ba19755467c5f08d3596567a1f41f37ed5fdf117db7e617b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.296282 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-k8ptd"] Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.297034 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:14 crc kubenswrapper[4998]: E1203 16:04:14.297129 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.302032 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65cf533-fd60-47a3-aab2-b7377bab9f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47k9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.319588 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.334248 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zh9r7\" (UniqueName: \"kubernetes.io/projected/2914eb52-522e-4789-a988-b15875755144-kube-api-access-zh9r7\") pod \"network-metrics-daemon-k8ptd\" (UID: \"2914eb52-522e-4789-a988-b15875755144\") " pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.334435 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2914eb52-522e-4789-a988-b15875755144-metrics-certs\") pod \"network-metrics-daemon-k8ptd\" (UID: \"2914eb52-522e-4789-a988-b15875755144\") " pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.350812 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"message\\\":\\\"ler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 16:04:11.823033 6464 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:04:11.823064 6464 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:04:11.823080 6464 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:04:11.823178 6464 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:11.823309 6464 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823440 6464 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823488 6464 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823628 6464 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:04:11.823784 6464 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:11.824254 6464 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-m2lh2_openshift-ovn-kubernetes(4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.377892 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.377961 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.377981 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.378008 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.378026 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:14Z","lastTransitionTime":"2025-12-03T16:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.385339 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.407932 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.427893 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.435091 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zh9r7\" (UniqueName: \"kubernetes.io/projected/2914eb52-522e-4789-a988-b15875755144-kube-api-access-zh9r7\") pod \"network-metrics-daemon-k8ptd\" (UID: \"2914eb52-522e-4789-a988-b15875755144\") " pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.435374 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2914eb52-522e-4789-a988-b15875755144-metrics-certs\") pod \"network-metrics-daemon-k8ptd\" (UID: \"2914eb52-522e-4789-a988-b15875755144\") " pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:14 crc kubenswrapper[4998]: E1203 16:04:14.435557 4998 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:04:14 crc kubenswrapper[4998]: E1203 16:04:14.435687 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2914eb52-522e-4789-a988-b15875755144-metrics-certs podName:2914eb52-522e-4789-a988-b15875755144 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:14.93565107 +0000 UTC m=+33.547351373 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2914eb52-522e-4789-a988-b15875755144-metrics-certs") pod "network-metrics-daemon-k8ptd" (UID: "2914eb52-522e-4789-a988-b15875755144") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.448401 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.464065 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.465680 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zh9r7\" (UniqueName: \"kubernetes.io/projected/2914eb52-522e-4789-a988-b15875755144-kube-api-access-zh9r7\") pod \"network-metrics-daemon-k8ptd\" (UID: \"2914eb52-522e-4789-a988-b15875755144\") " pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.481083 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.481137 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.481157 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.481184 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.481202 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:14Z","lastTransitionTime":"2025-12-03T16:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.489139 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.507155 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k8ptd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2914eb52-522e-4789-a988-b15875755144\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k8ptd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.529193 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.547357 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.566295 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d856098-4acb-4d57-8a27-de806955228c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53614cedb63f775d2b563e4405055cd6702c3f35980bc092ac07a3c9ff69aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c911081098fed822c164c7e59c1518ec11cf1c612e3ac4c84af8e694498c915d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9cdcf95d23503cce6d3fbc6de109783933aaaec4eefccd6f1834e234ed41f6e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.584549 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.584625 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.584648 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.584680 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.584703 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:14Z","lastTransitionTime":"2025-12-03T16:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.586333 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.613704 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e48502ef54b1ba19755467c5f08d3596567a1f41f37ed5fdf117db7e617b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.631374 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65cf533-fd60-47a3-aab2-b7377bab9f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44d20df9678d9388f9611bf182cf962781bbcb668bccef2fba40783ca6cd7cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93a183a13918fa4a6c42b6d0feed52f46b3eac00d85977936772b4cd7e4cb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47k9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.648872 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.660860 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:14Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.677294 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.677292 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:14 crc kubenswrapper[4998]: E1203 16:04:14.677464 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.677317 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:14 crc kubenswrapper[4998]: E1203 16:04:14.677546 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:14 crc kubenswrapper[4998]: E1203 16:04:14.677603 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.687597 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.687641 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.687660 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.687681 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.687698 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:14Z","lastTransitionTime":"2025-12-03T16:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.790162 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.790201 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.790211 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.790226 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.790238 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:14Z","lastTransitionTime":"2025-12-03T16:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.892944 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.892999 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.893011 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.893047 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.893062 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:14Z","lastTransitionTime":"2025-12-03T16:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.941998 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2914eb52-522e-4789-a988-b15875755144-metrics-certs\") pod \"network-metrics-daemon-k8ptd\" (UID: \"2914eb52-522e-4789-a988-b15875755144\") " pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:14 crc kubenswrapper[4998]: E1203 16:04:14.942184 4998 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:04:14 crc kubenswrapper[4998]: E1203 16:04:14.942260 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2914eb52-522e-4789-a988-b15875755144-metrics-certs podName:2914eb52-522e-4789-a988-b15875755144 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:15.942237594 +0000 UTC m=+34.553937857 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2914eb52-522e-4789-a988-b15875755144-metrics-certs") pod "network-metrics-daemon-k8ptd" (UID: "2914eb52-522e-4789-a988-b15875755144") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.995573 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.995624 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.995635 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.995653 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:14 crc kubenswrapper[4998]: I1203 16:04:14.995665 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:14Z","lastTransitionTime":"2025-12-03T16:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.098712 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.098786 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.098799 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.098820 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.098834 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:15Z","lastTransitionTime":"2025-12-03T16:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.202048 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.202122 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.202147 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.202180 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.202205 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:15Z","lastTransitionTime":"2025-12-03T16:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.305576 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.305615 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.305626 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.305644 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.305655 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:15Z","lastTransitionTime":"2025-12-03T16:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.408189 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.408242 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.408255 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.408274 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.408286 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:15Z","lastTransitionTime":"2025-12-03T16:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.433472 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.450533 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.467221 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.485140 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"message\\\":\\\"ler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 16:04:11.823033 6464 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:04:11.823064 6464 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:04:11.823080 6464 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:04:11.823178 6464 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:11.823309 6464 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823440 6464 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823488 6464 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823628 6464 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:04:11.823784 6464 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:11.824254 6464 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-m2lh2_openshift-ovn-kubernetes(4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.504594 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.510786 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.510810 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.510817 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.510831 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.510841 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:15Z","lastTransitionTime":"2025-12-03T16:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.519235 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.530553 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.543842 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.552351 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.562001 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.571717 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k8ptd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2914eb52-522e-4789-a988-b15875755144\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k8ptd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.582683 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.594455 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65cf533-fd60-47a3-aab2-b7377bab9f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44d20df9678d9388f9611bf182cf962781bbcb668bccef2fba40783ca6cd7cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93a183a13918fa4a6c42b6d0feed52f46b3eac00d85977936772b4cd7e4cb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47k9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.606181 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d856098-4acb-4d57-8a27-de806955228c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53614cedb63f775d2b563e4405055cd6702c3f35980bc092ac07a3c9ff69aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c911081098fed822c164c7e59c1518ec11cf1c612e3ac4c84af8e694498c915d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9cdcf95d23503cce6d3fbc6de109783933aaaec4eefccd6f1834e234ed41f6e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.613839 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.613874 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.613885 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.613901 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.613913 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:15Z","lastTransitionTime":"2025-12-03T16:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.618865 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.632941 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e48502ef54b1ba19755467c5f08d3596567a1f41f37ed5fdf117db7e617b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.646186 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.663255 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:15Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.677792 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:15 crc kubenswrapper[4998]: E1203 16:04:15.678044 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.717267 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.717315 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.717326 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.717342 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.717354 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:15Z","lastTransitionTime":"2025-12-03T16:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.820389 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.820456 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.820472 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.820494 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.820511 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:15Z","lastTransitionTime":"2025-12-03T16:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.923137 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.923234 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.923257 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.923279 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.923293 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:15Z","lastTransitionTime":"2025-12-03T16:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:15 crc kubenswrapper[4998]: I1203 16:04:15.951785 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2914eb52-522e-4789-a988-b15875755144-metrics-certs\") pod \"network-metrics-daemon-k8ptd\" (UID: \"2914eb52-522e-4789-a988-b15875755144\") " pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:15 crc kubenswrapper[4998]: E1203 16:04:15.951983 4998 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:04:15 crc kubenswrapper[4998]: E1203 16:04:15.952082 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2914eb52-522e-4789-a988-b15875755144-metrics-certs podName:2914eb52-522e-4789-a988-b15875755144 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:17.952059463 +0000 UTC m=+36.563759766 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2914eb52-522e-4789-a988-b15875755144-metrics-certs") pod "network-metrics-daemon-k8ptd" (UID: "2914eb52-522e-4789-a988-b15875755144") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.026126 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.026176 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.026193 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.026218 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.026235 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:16Z","lastTransitionTime":"2025-12-03T16:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.129454 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.129509 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.129527 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.129550 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.129567 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:16Z","lastTransitionTime":"2025-12-03T16:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.232911 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.232984 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.233006 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.233035 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.233056 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:16Z","lastTransitionTime":"2025-12-03T16:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.336218 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.336293 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.336312 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.336338 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.336357 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:16Z","lastTransitionTime":"2025-12-03T16:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.438563 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.438608 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.438620 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.438638 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.438650 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:16Z","lastTransitionTime":"2025-12-03T16:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.456105 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.456203 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.456228 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.456256 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.456280 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:16 crc kubenswrapper[4998]: E1203 16:04:16.456383 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:04:16 crc kubenswrapper[4998]: E1203 16:04:16.456401 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:04:16 crc kubenswrapper[4998]: E1203 16:04:16.456411 4998 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:16 crc kubenswrapper[4998]: E1203 16:04:16.456451 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:32.456438705 +0000 UTC m=+51.068138928 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:16 crc kubenswrapper[4998]: E1203 16:04:16.456488 4998 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:04:16 crc kubenswrapper[4998]: E1203 16:04:16.456509 4998 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:04:16 crc kubenswrapper[4998]: E1203 16:04:16.456560 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:32.456540557 +0000 UTC m=+51.068240840 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:04:16 crc kubenswrapper[4998]: E1203 16:04:16.456577 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:32.456569458 +0000 UTC m=+51.068269771 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:04:16 crc kubenswrapper[4998]: E1203 16:04:16.456594 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:04:32.456589458 +0000 UTC m=+51.068289681 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:04:16 crc kubenswrapper[4998]: E1203 16:04:16.456701 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:04:16 crc kubenswrapper[4998]: E1203 16:04:16.456721 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:04:16 crc kubenswrapper[4998]: E1203 16:04:16.456734 4998 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:16 crc kubenswrapper[4998]: E1203 16:04:16.456808 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:32.456796363 +0000 UTC m=+51.068496646 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.540687 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.540724 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.540733 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.540764 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.540776 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:16Z","lastTransitionTime":"2025-12-03T16:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.644371 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.644431 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.644453 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.644484 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.644506 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:16Z","lastTransitionTime":"2025-12-03T16:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.677169 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.677207 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.677175 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:16 crc kubenswrapper[4998]: E1203 16:04:16.677332 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:16 crc kubenswrapper[4998]: E1203 16:04:16.677496 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:16 crc kubenswrapper[4998]: E1203 16:04:16.677685 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.747504 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.747581 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.747597 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.747658 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.747677 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:16Z","lastTransitionTime":"2025-12-03T16:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.849895 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.849925 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.849934 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.849948 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.849957 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:16Z","lastTransitionTime":"2025-12-03T16:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.952978 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.953056 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.953079 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.953103 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:16 crc kubenswrapper[4998]: I1203 16:04:16.953119 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:16Z","lastTransitionTime":"2025-12-03T16:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.056120 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.056156 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.056167 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.056183 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.056195 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:17Z","lastTransitionTime":"2025-12-03T16:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.159545 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.159613 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.159632 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.159657 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.159673 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:17Z","lastTransitionTime":"2025-12-03T16:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.263919 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.263973 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.263984 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.264003 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.264014 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:17Z","lastTransitionTime":"2025-12-03T16:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.366472 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.366527 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.366542 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.366563 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.366579 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:17Z","lastTransitionTime":"2025-12-03T16:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.469222 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.469283 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.469299 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.469323 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.469340 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:17Z","lastTransitionTime":"2025-12-03T16:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.572431 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.572473 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.572485 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.572502 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.572512 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:17Z","lastTransitionTime":"2025-12-03T16:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.675510 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.675550 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.675561 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.675578 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.675588 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:17Z","lastTransitionTime":"2025-12-03T16:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.676948 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:17 crc kubenswrapper[4998]: E1203 16:04:17.677055 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.779031 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.779072 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.779086 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.779103 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.779118 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:17Z","lastTransitionTime":"2025-12-03T16:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.881519 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.881559 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.881569 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.881586 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.881599 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:17Z","lastTransitionTime":"2025-12-03T16:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.974002 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2914eb52-522e-4789-a988-b15875755144-metrics-certs\") pod \"network-metrics-daemon-k8ptd\" (UID: \"2914eb52-522e-4789-a988-b15875755144\") " pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:17 crc kubenswrapper[4998]: E1203 16:04:17.974141 4998 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:04:17 crc kubenswrapper[4998]: E1203 16:04:17.974203 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2914eb52-522e-4789-a988-b15875755144-metrics-certs podName:2914eb52-522e-4789-a988-b15875755144 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:21.974185848 +0000 UTC m=+40.585886081 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2914eb52-522e-4789-a988-b15875755144-metrics-certs") pod "network-metrics-daemon-k8ptd" (UID: "2914eb52-522e-4789-a988-b15875755144") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.984014 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.984050 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.984061 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.984083 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:17 crc kubenswrapper[4998]: I1203 16:04:17.984099 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:17Z","lastTransitionTime":"2025-12-03T16:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.086144 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.086188 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.086201 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.086221 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.086234 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:18Z","lastTransitionTime":"2025-12-03T16:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.189030 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.189103 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.189121 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.189147 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.189167 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:18Z","lastTransitionTime":"2025-12-03T16:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.291043 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.291102 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.291119 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.291147 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.291218 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:18Z","lastTransitionTime":"2025-12-03T16:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.394157 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.394220 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.394242 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.394270 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.394291 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:18Z","lastTransitionTime":"2025-12-03T16:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.497343 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.497416 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.497432 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.497455 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.497472 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:18Z","lastTransitionTime":"2025-12-03T16:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.600674 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.600728 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.600742 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.600788 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.600802 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:18Z","lastTransitionTime":"2025-12-03T16:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.677670 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.677741 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:18 crc kubenswrapper[4998]: E1203 16:04:18.677811 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:18 crc kubenswrapper[4998]: E1203 16:04:18.677926 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.677781 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:18 crc kubenswrapper[4998]: E1203 16:04:18.678126 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.704197 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.704258 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.704274 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.704299 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.704316 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:18Z","lastTransitionTime":"2025-12-03T16:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.807862 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.807900 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.807912 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.807938 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.807950 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:18Z","lastTransitionTime":"2025-12-03T16:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.910560 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.910634 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.910658 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.910690 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:18 crc kubenswrapper[4998]: I1203 16:04:18.910713 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:18Z","lastTransitionTime":"2025-12-03T16:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.013236 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.013285 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.013296 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.013315 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.013329 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:19Z","lastTransitionTime":"2025-12-03T16:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.115983 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.116023 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.116035 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.116052 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.116066 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:19Z","lastTransitionTime":"2025-12-03T16:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.219022 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.219081 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.219098 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.219121 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.219138 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:19Z","lastTransitionTime":"2025-12-03T16:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.322081 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.322125 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.322137 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.322155 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.322168 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:19Z","lastTransitionTime":"2025-12-03T16:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.424843 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.424917 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.424940 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.424974 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.424998 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:19Z","lastTransitionTime":"2025-12-03T16:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.527962 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.528011 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.528022 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.528039 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.528050 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:19Z","lastTransitionTime":"2025-12-03T16:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.630209 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.630284 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.630309 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.630334 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.630351 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:19Z","lastTransitionTime":"2025-12-03T16:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.677328 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:19 crc kubenswrapper[4998]: E1203 16:04:19.677511 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.732622 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.732675 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.732683 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.732698 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.732709 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:19Z","lastTransitionTime":"2025-12-03T16:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.835932 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.835972 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.835983 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.835999 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.836011 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:19Z","lastTransitionTime":"2025-12-03T16:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.938860 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.938930 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.938950 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.938974 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:19 crc kubenswrapper[4998]: I1203 16:04:19.938991 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:19Z","lastTransitionTime":"2025-12-03T16:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.041456 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.041510 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.041529 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.041554 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.041574 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:20Z","lastTransitionTime":"2025-12-03T16:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.145214 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.145281 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.145302 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.145331 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.145351 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:20Z","lastTransitionTime":"2025-12-03T16:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.248654 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.248717 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.248738 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.248805 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.248832 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:20Z","lastTransitionTime":"2025-12-03T16:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.352113 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.352185 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.352201 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.352226 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.352242 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:20Z","lastTransitionTime":"2025-12-03T16:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.455451 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.455524 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.455544 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.455570 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.455588 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:20Z","lastTransitionTime":"2025-12-03T16:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.558244 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.558298 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.558316 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.558340 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.558358 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:20Z","lastTransitionTime":"2025-12-03T16:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.637434 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.637512 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.637525 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.637543 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.637554 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:20Z","lastTransitionTime":"2025-12-03T16:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:20 crc kubenswrapper[4998]: E1203 16:04:20.655538 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:20Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.661178 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.661270 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.661288 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.661340 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.661367 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:20Z","lastTransitionTime":"2025-12-03T16:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.677700 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.677785 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.677712 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:20 crc kubenswrapper[4998]: E1203 16:04:20.677991 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:20 crc kubenswrapper[4998]: E1203 16:04:20.678105 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:20 crc kubenswrapper[4998]: E1203 16:04:20.678283 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:20 crc kubenswrapper[4998]: E1203 16:04:20.678951 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:20Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.683237 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.683269 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.683278 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.683290 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.683299 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:20Z","lastTransitionTime":"2025-12-03T16:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:20 crc kubenswrapper[4998]: E1203 16:04:20.697959 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:20Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.702313 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.702379 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.702390 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.702429 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.702441 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:20Z","lastTransitionTime":"2025-12-03T16:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:20 crc kubenswrapper[4998]: E1203 16:04:20.722490 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:20Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.727358 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.727436 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.727460 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.727493 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.727510 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:20Z","lastTransitionTime":"2025-12-03T16:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:20 crc kubenswrapper[4998]: E1203 16:04:20.747386 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:20Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:20 crc kubenswrapper[4998]: E1203 16:04:20.747614 4998 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.749888 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.749927 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.749939 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.749955 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.749966 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:20Z","lastTransitionTime":"2025-12-03T16:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.852569 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.852650 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.852673 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.852696 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.852714 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:20Z","lastTransitionTime":"2025-12-03T16:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.955721 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.955869 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.955889 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.955912 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:20 crc kubenswrapper[4998]: I1203 16:04:20.955961 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:20Z","lastTransitionTime":"2025-12-03T16:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.059011 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.059062 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.059074 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.059091 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.059105 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:21Z","lastTransitionTime":"2025-12-03T16:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.162171 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.162232 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.162248 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.162276 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.162293 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:21Z","lastTransitionTime":"2025-12-03T16:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.265694 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.265781 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.265801 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.265825 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.265842 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:21Z","lastTransitionTime":"2025-12-03T16:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.369091 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.369148 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.369160 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.369179 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.369192 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:21Z","lastTransitionTime":"2025-12-03T16:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.471899 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.471969 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.471985 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.472010 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.472027 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:21Z","lastTransitionTime":"2025-12-03T16:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.574547 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.574597 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.574647 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.574665 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.574682 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:21Z","lastTransitionTime":"2025-12-03T16:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.677277 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:21 crc kubenswrapper[4998]: E1203 16:04:21.677423 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.677518 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.677575 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.677613 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.677639 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.677657 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:21Z","lastTransitionTime":"2025-12-03T16:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.700826 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:21Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.720503 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:21Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.741994 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:21Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.757406 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:21Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.775929 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k8ptd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2914eb52-522e-4789-a988-b15875755144\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k8ptd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:21Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.779461 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.779516 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.779530 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.779550 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.779566 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:21Z","lastTransitionTime":"2025-12-03T16:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.794320 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d856098-4acb-4d57-8a27-de806955228c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53614cedb63f775d2b563e4405055cd6702c3f35980bc092ac07a3c9ff69aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c911081098fed822c164c7e59c1518ec11cf1c612e3ac4c84af8e694498c915d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9cdcf95d23503cce6d3fbc6de109783933aaaec4eefccd6f1834e234ed41f6e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:21Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.813689 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:21Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.835420 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e48502ef54b1ba19755467c5f08d3596567a1f41f37ed5fdf117db7e617b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:21Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.851337 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65cf533-fd60-47a3-aab2-b7377bab9f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44d20df9678d9388f9611bf182cf962781bbcb668bccef2fba40783ca6cd7cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93a183a13918fa4a6c42b6d0feed52f46b3eac00d85977936772b4cd7e4cb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47k9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:21Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.865848 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:21Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.882454 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.882511 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.882528 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.882554 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.882573 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:21Z","lastTransitionTime":"2025-12-03T16:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.885415 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:21Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.916520 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:21Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.934187 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:21Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.949075 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:21Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.963690 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:21Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.977848 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:21Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.985114 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.985180 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.985207 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.985233 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:21 crc kubenswrapper[4998]: I1203 16:04:21.985251 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:21Z","lastTransitionTime":"2025-12-03T16:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.015317 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"message\\\":\\\"ler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 16:04:11.823033 6464 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:04:11.823064 6464 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:04:11.823080 6464 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:04:11.823178 6464 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:11.823309 6464 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823440 6464 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823488 6464 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823628 6464 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:04:11.823784 6464 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:11.824254 6464 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-m2lh2_openshift-ovn-kubernetes(4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:22Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.018636 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2914eb52-522e-4789-a988-b15875755144-metrics-certs\") pod \"network-metrics-daemon-k8ptd\" (UID: \"2914eb52-522e-4789-a988-b15875755144\") " pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:22 crc kubenswrapper[4998]: E1203 16:04:22.018916 4998 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:04:22 crc kubenswrapper[4998]: E1203 16:04:22.019050 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2914eb52-522e-4789-a988-b15875755144-metrics-certs podName:2914eb52-522e-4789-a988-b15875755144 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:30.019014561 +0000 UTC m=+48.630714864 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2914eb52-522e-4789-a988-b15875755144-metrics-certs") pod "network-metrics-daemon-k8ptd" (UID: "2914eb52-522e-4789-a988-b15875755144") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.088057 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.088112 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.088131 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.088155 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.088175 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:22Z","lastTransitionTime":"2025-12-03T16:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.096315 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.097566 4998 scope.go:117] "RemoveContainer" containerID="609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95" Dec 03 16:04:22 crc kubenswrapper[4998]: E1203 16:04:22.097883 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-m2lh2_openshift-ovn-kubernetes(4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.191338 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.191406 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.191429 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.191459 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.191480 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:22Z","lastTransitionTime":"2025-12-03T16:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.294058 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.294116 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.294135 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.294160 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.294179 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:22Z","lastTransitionTime":"2025-12-03T16:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.397714 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.397809 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.397827 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.397852 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.397869 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:22Z","lastTransitionTime":"2025-12-03T16:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.500715 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.500811 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.500831 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.500856 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.500872 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:22Z","lastTransitionTime":"2025-12-03T16:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.604645 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.604730 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.604789 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.604824 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.604847 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:22Z","lastTransitionTime":"2025-12-03T16:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.677667 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.677787 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:22 crc kubenswrapper[4998]: E1203 16:04:22.677935 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.677672 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:22 crc kubenswrapper[4998]: E1203 16:04:22.678135 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:22 crc kubenswrapper[4998]: E1203 16:04:22.678271 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.706937 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.706986 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.707004 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.707028 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.707045 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:22Z","lastTransitionTime":"2025-12-03T16:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.810159 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.810563 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.810715 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.810921 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.811088 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:22Z","lastTransitionTime":"2025-12-03T16:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.914805 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.914854 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.914866 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.914902 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:22 crc kubenswrapper[4998]: I1203 16:04:22.914915 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:22Z","lastTransitionTime":"2025-12-03T16:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.017838 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.017893 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.017916 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.017943 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.017965 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:23Z","lastTransitionTime":"2025-12-03T16:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.120164 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.120257 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.120284 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.120314 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.120337 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:23Z","lastTransitionTime":"2025-12-03T16:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.223686 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.223747 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.223824 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.223855 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.223877 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:23Z","lastTransitionTime":"2025-12-03T16:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.326011 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.326045 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.326053 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.326066 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.326075 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:23Z","lastTransitionTime":"2025-12-03T16:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.429032 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.429087 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.429104 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.429130 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.429146 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:23Z","lastTransitionTime":"2025-12-03T16:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.533177 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.533255 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.533279 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.533309 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.533330 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:23Z","lastTransitionTime":"2025-12-03T16:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.636568 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.636651 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.636679 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.636709 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.636732 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:23Z","lastTransitionTime":"2025-12-03T16:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.677210 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:23 crc kubenswrapper[4998]: E1203 16:04:23.677412 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.738660 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.738727 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.738744 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.738806 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.738829 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:23Z","lastTransitionTime":"2025-12-03T16:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.842201 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.842272 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.842297 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.842329 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.842353 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:23Z","lastTransitionTime":"2025-12-03T16:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.945483 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.945539 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.945554 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.945575 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:23 crc kubenswrapper[4998]: I1203 16:04:23.945589 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:23Z","lastTransitionTime":"2025-12-03T16:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.048780 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.048845 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.048858 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.048879 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.048892 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:24Z","lastTransitionTime":"2025-12-03T16:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.152193 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.152323 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.152343 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.152373 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.152391 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:24Z","lastTransitionTime":"2025-12-03T16:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.255661 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.255724 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.255745 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.255791 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.255807 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:24Z","lastTransitionTime":"2025-12-03T16:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.359507 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.359587 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.359606 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.359635 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.359654 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:24Z","lastTransitionTime":"2025-12-03T16:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.463149 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.463231 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.463261 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.463295 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.463317 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:24Z","lastTransitionTime":"2025-12-03T16:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.566280 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.566371 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.566398 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.566438 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.566465 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:24Z","lastTransitionTime":"2025-12-03T16:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.670359 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.670437 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.670462 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.670512 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.670535 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:24Z","lastTransitionTime":"2025-12-03T16:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.677657 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.677722 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.677834 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:24 crc kubenswrapper[4998]: E1203 16:04:24.677940 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:24 crc kubenswrapper[4998]: E1203 16:04:24.678096 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:24 crc kubenswrapper[4998]: E1203 16:04:24.678261 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.773954 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.774030 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.774049 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.774080 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.774103 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:24Z","lastTransitionTime":"2025-12-03T16:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.876607 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.876647 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.876659 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.876675 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.876686 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:24Z","lastTransitionTime":"2025-12-03T16:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.979796 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.979864 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.979883 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.979909 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:24 crc kubenswrapper[4998]: I1203 16:04:24.979927 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:24Z","lastTransitionTime":"2025-12-03T16:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.083873 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.084253 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.084341 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.084464 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.084617 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:25Z","lastTransitionTime":"2025-12-03T16:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.187844 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.187907 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.187922 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.187945 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.187959 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:25Z","lastTransitionTime":"2025-12-03T16:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.290696 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.290727 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.290735 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.290747 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.290775 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:25Z","lastTransitionTime":"2025-12-03T16:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.393381 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.393432 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.393445 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.393464 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.393477 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:25Z","lastTransitionTime":"2025-12-03T16:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.496503 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.496600 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.496622 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.496649 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.496669 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:25Z","lastTransitionTime":"2025-12-03T16:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.599217 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.599251 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.599259 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.599272 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.599281 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:25Z","lastTransitionTime":"2025-12-03T16:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.677632 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:25 crc kubenswrapper[4998]: E1203 16:04:25.677734 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.701634 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.701689 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.701701 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.701717 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.701732 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:25Z","lastTransitionTime":"2025-12-03T16:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.805557 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.805611 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.805624 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.805643 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.805657 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:25Z","lastTransitionTime":"2025-12-03T16:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.908146 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.908209 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.908231 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.908292 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:25 crc kubenswrapper[4998]: I1203 16:04:25.908318 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:25Z","lastTransitionTime":"2025-12-03T16:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.010528 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.010600 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.010622 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.010652 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.010674 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:26Z","lastTransitionTime":"2025-12-03T16:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.113604 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.113654 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.113669 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.113687 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.113699 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:26Z","lastTransitionTime":"2025-12-03T16:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.216600 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.216675 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.216697 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.216730 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.216793 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:26Z","lastTransitionTime":"2025-12-03T16:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.319620 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.319678 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.319693 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.319714 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.319725 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:26Z","lastTransitionTime":"2025-12-03T16:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.422749 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.422820 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.422836 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.422855 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.422866 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:26Z","lastTransitionTime":"2025-12-03T16:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.525351 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.525413 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.525433 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.525456 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.525475 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:26Z","lastTransitionTime":"2025-12-03T16:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.627572 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.627612 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.627623 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.627640 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.627650 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:26Z","lastTransitionTime":"2025-12-03T16:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.677805 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.677927 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.677813 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:26 crc kubenswrapper[4998]: E1203 16:04:26.677987 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:26 crc kubenswrapper[4998]: E1203 16:04:26.678112 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:26 crc kubenswrapper[4998]: E1203 16:04:26.678237 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.730842 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.730932 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.730966 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.731000 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.731022 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:26Z","lastTransitionTime":"2025-12-03T16:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.833085 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.833148 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.833166 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.833193 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.833211 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:26Z","lastTransitionTime":"2025-12-03T16:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.937365 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.937475 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.937502 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.937594 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:26 crc kubenswrapper[4998]: I1203 16:04:26.937662 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:26Z","lastTransitionTime":"2025-12-03T16:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.041263 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.041809 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.041886 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.041947 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.042010 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:27Z","lastTransitionTime":"2025-12-03T16:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.144880 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.144952 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.144987 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.145031 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.145054 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:27Z","lastTransitionTime":"2025-12-03T16:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.247803 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.247917 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.247957 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.247990 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.248011 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:27Z","lastTransitionTime":"2025-12-03T16:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.351547 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.351620 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.351640 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.351666 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.351688 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:27Z","lastTransitionTime":"2025-12-03T16:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.455407 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.455471 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.455489 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.455513 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.455544 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:27Z","lastTransitionTime":"2025-12-03T16:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.559358 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.559442 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.559463 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.559493 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.559514 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:27Z","lastTransitionTime":"2025-12-03T16:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.662724 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.662833 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.662852 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.662879 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.662898 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:27Z","lastTransitionTime":"2025-12-03T16:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.677241 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:27 crc kubenswrapper[4998]: E1203 16:04:27.677445 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.765873 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.765959 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.765985 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.766020 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.766045 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:27Z","lastTransitionTime":"2025-12-03T16:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.869586 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.869672 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.869702 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.869738 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.869793 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:27Z","lastTransitionTime":"2025-12-03T16:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.973914 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.973976 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.973994 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.974024 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:27 crc kubenswrapper[4998]: I1203 16:04:27.974041 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:27Z","lastTransitionTime":"2025-12-03T16:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.077194 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.078003 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.078076 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.078112 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.078147 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:28Z","lastTransitionTime":"2025-12-03T16:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.181123 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.181194 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.181215 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.181245 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.181264 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:28Z","lastTransitionTime":"2025-12-03T16:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.284666 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.284727 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.284746 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.284798 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.284818 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:28Z","lastTransitionTime":"2025-12-03T16:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.388662 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.388740 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.388799 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.388835 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.388860 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:28Z","lastTransitionTime":"2025-12-03T16:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.492147 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.492216 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.492231 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.492258 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.492272 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:28Z","lastTransitionTime":"2025-12-03T16:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.595826 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.595888 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.595910 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.595937 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.595956 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:28Z","lastTransitionTime":"2025-12-03T16:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.677406 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.677519 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:28 crc kubenswrapper[4998]: E1203 16:04:28.677602 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.677514 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:28 crc kubenswrapper[4998]: E1203 16:04:28.677706 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:28 crc kubenswrapper[4998]: E1203 16:04:28.677938 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.698894 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.698942 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.698953 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.698975 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.698989 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:28Z","lastTransitionTime":"2025-12-03T16:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.801577 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.801644 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.801655 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.801682 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.801696 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:28Z","lastTransitionTime":"2025-12-03T16:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.904153 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.904219 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.904234 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.904258 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:28 crc kubenswrapper[4998]: I1203 16:04:28.904274 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:28Z","lastTransitionTime":"2025-12-03T16:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.007236 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.007315 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.007332 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.007359 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.007383 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:29Z","lastTransitionTime":"2025-12-03T16:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.109850 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.109922 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.109939 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.109966 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.109983 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:29Z","lastTransitionTime":"2025-12-03T16:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.212284 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.212339 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.212351 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.212367 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.212379 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:29Z","lastTransitionTime":"2025-12-03T16:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.316161 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.316256 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.316278 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.316306 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.316335 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:29Z","lastTransitionTime":"2025-12-03T16:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.419251 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.419315 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.419333 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.419358 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.419374 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:29Z","lastTransitionTime":"2025-12-03T16:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.522385 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.522446 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.522463 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.522489 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.522506 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:29Z","lastTransitionTime":"2025-12-03T16:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.624972 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.625025 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.625040 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.625059 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.625073 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:29Z","lastTransitionTime":"2025-12-03T16:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.676901 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:29 crc kubenswrapper[4998]: E1203 16:04:29.677113 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.728352 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.728411 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.728429 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.728457 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.728475 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:29Z","lastTransitionTime":"2025-12-03T16:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.831459 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.831515 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.831556 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.831569 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.831579 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:29Z","lastTransitionTime":"2025-12-03T16:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.933858 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.933903 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.933920 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.933943 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:29 crc kubenswrapper[4998]: I1203 16:04:29.933960 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:29Z","lastTransitionTime":"2025-12-03T16:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.035944 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.036003 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.036014 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.036033 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.036043 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:30Z","lastTransitionTime":"2025-12-03T16:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.112173 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2914eb52-522e-4789-a988-b15875755144-metrics-certs\") pod \"network-metrics-daemon-k8ptd\" (UID: \"2914eb52-522e-4789-a988-b15875755144\") " pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:30 crc kubenswrapper[4998]: E1203 16:04:30.112322 4998 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:04:30 crc kubenswrapper[4998]: E1203 16:04:30.112415 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2914eb52-522e-4789-a988-b15875755144-metrics-certs podName:2914eb52-522e-4789-a988-b15875755144 nodeName:}" failed. No retries permitted until 2025-12-03 16:04:46.112397071 +0000 UTC m=+64.724097294 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2914eb52-522e-4789-a988-b15875755144-metrics-certs") pod "network-metrics-daemon-k8ptd" (UID: "2914eb52-522e-4789-a988-b15875755144") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.138958 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.139015 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.139033 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.139060 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.139080 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:30Z","lastTransitionTime":"2025-12-03T16:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.242408 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.242484 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.242505 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.242530 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.242548 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:30Z","lastTransitionTime":"2025-12-03T16:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.345697 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.345801 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.345827 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.345863 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.345882 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:30Z","lastTransitionTime":"2025-12-03T16:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.448990 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.449059 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.449083 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.449112 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.449145 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:30Z","lastTransitionTime":"2025-12-03T16:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.552265 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.552322 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.552339 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.552364 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.552381 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:30Z","lastTransitionTime":"2025-12-03T16:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.655313 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.655358 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.655380 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.655426 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.655450 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:30Z","lastTransitionTime":"2025-12-03T16:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.677321 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.677377 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.677343 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:30 crc kubenswrapper[4998]: E1203 16:04:30.677520 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:30 crc kubenswrapper[4998]: E1203 16:04:30.677688 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:30 crc kubenswrapper[4998]: E1203 16:04:30.677877 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.759281 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.759587 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.759808 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.759971 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.760152 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:30Z","lastTransitionTime":"2025-12-03T16:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.827526 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.827883 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.827999 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.828134 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.828247 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:30Z","lastTransitionTime":"2025-12-03T16:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:30 crc kubenswrapper[4998]: E1203 16:04:30.850658 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:30Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.856524 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.856595 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.856613 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.856638 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.856656 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:30Z","lastTransitionTime":"2025-12-03T16:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:30 crc kubenswrapper[4998]: E1203 16:04:30.891079 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:30Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.898959 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.899020 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.899039 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.899065 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.899084 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:30Z","lastTransitionTime":"2025-12-03T16:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:30 crc kubenswrapper[4998]: E1203 16:04:30.927099 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:30Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.933746 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.933999 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.934119 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.934233 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.934350 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:30Z","lastTransitionTime":"2025-12-03T16:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:30 crc kubenswrapper[4998]: E1203 16:04:30.947185 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:30Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.951160 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.951193 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.951205 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.951223 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.951235 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:30Z","lastTransitionTime":"2025-12-03T16:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:30 crc kubenswrapper[4998]: E1203 16:04:30.963150 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:30Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:30 crc kubenswrapper[4998]: E1203 16:04:30.963362 4998 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.965055 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.965091 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.965106 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.965124 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:30 crc kubenswrapper[4998]: I1203 16:04:30.965136 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:30Z","lastTransitionTime":"2025-12-03T16:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.069947 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.070024 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.070078 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.070113 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.070134 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:31Z","lastTransitionTime":"2025-12-03T16:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.172959 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.173015 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.173033 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.173060 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.173078 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:31Z","lastTransitionTime":"2025-12-03T16:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.276522 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.276567 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.276584 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.276605 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.276620 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:31Z","lastTransitionTime":"2025-12-03T16:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.380025 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.380076 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.380092 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.380115 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.380132 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:31Z","lastTransitionTime":"2025-12-03T16:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.483216 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.483267 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.483289 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.483313 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.483331 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:31Z","lastTransitionTime":"2025-12-03T16:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.586085 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.586150 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.586175 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.586204 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.586225 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:31Z","lastTransitionTime":"2025-12-03T16:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.677247 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:31 crc kubenswrapper[4998]: E1203 16:04:31.677478 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.688715 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.688802 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.688820 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.688842 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.688861 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:31Z","lastTransitionTime":"2025-12-03T16:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.700910 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:31Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.722090 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:31Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.739918 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:31Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.757597 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:31Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.778803 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k8ptd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2914eb52-522e-4789-a988-b15875755144\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k8ptd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:31Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.791956 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.791999 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.792011 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.792026 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.792039 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:31Z","lastTransitionTime":"2025-12-03T16:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.798141 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:31Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.815565 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e48502ef54b1ba19755467c5f08d3596567a1f41f37ed5fdf117db7e617b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:31Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.828519 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65cf533-fd60-47a3-aab2-b7377bab9f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44d20df9678d9388f9611bf182cf962781bbcb668bccef2fba40783ca6cd7cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93a183a13918fa4a6c42b6d0feed52f46b3eac00d85977936772b4cd7e4cb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47k9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:31Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.844854 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d856098-4acb-4d57-8a27-de806955228c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53614cedb63f775d2b563e4405055cd6702c3f35980bc092ac07a3c9ff69aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c911081098fed822c164c7e59c1518ec11cf1c612e3ac4c84af8e694498c915d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9cdcf95d23503cce6d3fbc6de109783933aaaec4eefccd6f1834e234ed41f6e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:31Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.862328 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:31Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.877662 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:31Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.894683 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.894724 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.894739 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.894784 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.894802 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:31Z","lastTransitionTime":"2025-12-03T16:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.898156 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:31Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.916711 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:31Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.932109 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:31Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.951456 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:31Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.979179 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"message\\\":\\\"ler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 16:04:11.823033 6464 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:04:11.823064 6464 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:04:11.823080 6464 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:04:11.823178 6464 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:11.823309 6464 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823440 6464 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823488 6464 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823628 6464 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:04:11.823784 6464 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:11.824254 6464 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-m2lh2_openshift-ovn-kubernetes(4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:31Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.997181 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.997236 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.997256 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.997280 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:31 crc kubenswrapper[4998]: I1203 16:04:31.997298 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:31Z","lastTransitionTime":"2025-12-03T16:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.006653 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:32Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.100565 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.100623 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.100639 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.100658 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.100674 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:32Z","lastTransitionTime":"2025-12-03T16:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.203293 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.203363 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.203382 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.203410 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.203457 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:32Z","lastTransitionTime":"2025-12-03T16:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.306608 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.306644 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.306656 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.306672 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.306684 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:32Z","lastTransitionTime":"2025-12-03T16:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.409530 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.409570 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.409581 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.409600 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.409616 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:32Z","lastTransitionTime":"2025-12-03T16:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.513272 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.513326 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.513344 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.513367 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.513384 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:32Z","lastTransitionTime":"2025-12-03T16:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.540096 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:04:32 crc kubenswrapper[4998]: E1203 16:04:32.540269 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:04.540234899 +0000 UTC m=+83.151935162 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.540331 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.540418 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:32 crc kubenswrapper[4998]: E1203 16:04:32.540588 4998 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.540604 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:32 crc kubenswrapper[4998]: E1203 16:04:32.540658 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:05:04.540640569 +0000 UTC m=+83.152340802 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.540694 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:32 crc kubenswrapper[4998]: E1203 16:04:32.540900 4998 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:04:32 crc kubenswrapper[4998]: E1203 16:04:32.540945 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:05:04.540934055 +0000 UTC m=+83.152634288 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:04:32 crc kubenswrapper[4998]: E1203 16:04:32.540976 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:04:32 crc kubenswrapper[4998]: E1203 16:04:32.541008 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:04:32 crc kubenswrapper[4998]: E1203 16:04:32.541030 4998 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:32 crc kubenswrapper[4998]: E1203 16:04:32.541098 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:04:32 crc kubenswrapper[4998]: E1203 16:04:32.541147 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:04:32 crc kubenswrapper[4998]: E1203 16:04:32.541166 4998 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:32 crc kubenswrapper[4998]: E1203 16:04:32.541117 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 16:05:04.541096049 +0000 UTC m=+83.152796392 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:32 crc kubenswrapper[4998]: E1203 16:04:32.541260 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 16:05:04.541236073 +0000 UTC m=+83.152936376 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.616219 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.616283 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.616300 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.616324 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.616338 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:32Z","lastTransitionTime":"2025-12-03T16:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.677319 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.677448 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.677348 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:32 crc kubenswrapper[4998]: E1203 16:04:32.677516 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:32 crc kubenswrapper[4998]: E1203 16:04:32.677616 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:32 crc kubenswrapper[4998]: E1203 16:04:32.677721 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.719477 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.719542 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.719551 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.719565 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.719575 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:32Z","lastTransitionTime":"2025-12-03T16:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.821700 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.821744 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.821770 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.821789 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.821813 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:32Z","lastTransitionTime":"2025-12-03T16:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.924472 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.924564 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.924578 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.924595 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.924606 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:32Z","lastTransitionTime":"2025-12-03T16:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:32 crc kubenswrapper[4998]: I1203 16:04:32.989954 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.001998 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.011297 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d856098-4acb-4d57-8a27-de806955228c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53614cedb63f775d2b563e4405055cd6702c3f35980bc092ac07a3c9ff69aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c911081098fed822c164c7e59c1518ec11cf1c612e3ac4c84af8e694498c915d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9cdcf95d23503cce6d3fbc6de109783933aaaec4eefccd6f1834e234ed41f6e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.027785 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.027825 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.027837 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.027852 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.027862 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:33Z","lastTransitionTime":"2025-12-03T16:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.032242 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.055355 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e48502ef54b1ba19755467c5f08d3596567a1f41f37ed5fdf117db7e617b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.069105 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65cf533-fd60-47a3-aab2-b7377bab9f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44d20df9678d9388f9611bf182cf962781bbcb668bccef2fba40783ca6cd7cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93a183a13918fa4a6c42b6d0feed52f46b3eac00d85977936772b4cd7e4cb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47k9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.082967 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.098422 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.119372 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.130197 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.130407 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.130490 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.130574 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.130652 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:33Z","lastTransitionTime":"2025-12-03T16:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.134526 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.151028 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.164659 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.179686 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.200798 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"message\\\":\\\"ler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 16:04:11.823033 6464 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:04:11.823064 6464 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:04:11.823080 6464 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:04:11.823178 6464 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:11.823309 6464 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823440 6464 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823488 6464 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823628 6464 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:04:11.823784 6464 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:11.824254 6464 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-m2lh2_openshift-ovn-kubernetes(4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.218529 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.230905 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.232994 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.233033 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.233047 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.233072 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.233087 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:33Z","lastTransitionTime":"2025-12-03T16:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.243555 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.257871 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.272073 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k8ptd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2914eb52-522e-4789-a988-b15875755144\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k8ptd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:33Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.336553 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.336609 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.336623 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.336642 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.336654 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:33Z","lastTransitionTime":"2025-12-03T16:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.439735 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.440226 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.440254 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.440288 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.440320 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:33Z","lastTransitionTime":"2025-12-03T16:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.544281 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.544334 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.544345 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.544365 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.544377 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:33Z","lastTransitionTime":"2025-12-03T16:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.648173 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.648224 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.648242 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.648265 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.648283 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:33Z","lastTransitionTime":"2025-12-03T16:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.677939 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:33 crc kubenswrapper[4998]: E1203 16:04:33.678159 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.751042 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.751097 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.751113 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.751137 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.751157 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:33Z","lastTransitionTime":"2025-12-03T16:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.853819 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.853890 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.853914 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.853945 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.853970 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:33Z","lastTransitionTime":"2025-12-03T16:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.970689 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.970827 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.970852 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.970883 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:33 crc kubenswrapper[4998]: I1203 16:04:33.970910 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:33Z","lastTransitionTime":"2025-12-03T16:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.074220 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.074283 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.074308 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.074339 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.074364 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:34Z","lastTransitionTime":"2025-12-03T16:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.177660 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.177736 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.177782 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.177808 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.177828 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:34Z","lastTransitionTime":"2025-12-03T16:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.281192 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.281241 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.281253 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.281270 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.281286 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:34Z","lastTransitionTime":"2025-12-03T16:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.384678 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.384740 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.384789 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.384818 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.384836 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:34Z","lastTransitionTime":"2025-12-03T16:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.487252 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.487295 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.487308 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.487325 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.487341 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:34Z","lastTransitionTime":"2025-12-03T16:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.590212 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.590268 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.590285 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.590312 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.590329 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:34Z","lastTransitionTime":"2025-12-03T16:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.677572 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.677678 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.677588 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:34 crc kubenswrapper[4998]: E1203 16:04:34.677747 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:34 crc kubenswrapper[4998]: E1203 16:04:34.677902 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:34 crc kubenswrapper[4998]: E1203 16:04:34.678124 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.693095 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.693154 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.693173 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.693198 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.693221 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:34Z","lastTransitionTime":"2025-12-03T16:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.796638 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.796683 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.796695 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.796716 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.796744 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:34Z","lastTransitionTime":"2025-12-03T16:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.900685 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.900748 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.900798 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.900825 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:34 crc kubenswrapper[4998]: I1203 16:04:34.900843 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:34Z","lastTransitionTime":"2025-12-03T16:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.004318 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.004378 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.004394 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.004416 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.004435 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:35Z","lastTransitionTime":"2025-12-03T16:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.108221 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.108289 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.108311 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.108336 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.108359 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:35Z","lastTransitionTime":"2025-12-03T16:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.210892 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.210924 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.210933 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.210947 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.210958 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:35Z","lastTransitionTime":"2025-12-03T16:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.314255 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.314331 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.314350 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.314377 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.314395 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:35Z","lastTransitionTime":"2025-12-03T16:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.416579 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.416654 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.416663 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.416680 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.416698 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:35Z","lastTransitionTime":"2025-12-03T16:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.519709 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.519785 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.519800 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.519822 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.519837 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:35Z","lastTransitionTime":"2025-12-03T16:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.622931 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.622984 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.622997 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.623015 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.623028 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:35Z","lastTransitionTime":"2025-12-03T16:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.678117 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:35 crc kubenswrapper[4998]: E1203 16:04:35.678278 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.679668 4998 scope.go:117] "RemoveContainer" containerID="609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.725901 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.726231 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.726451 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.726681 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.726981 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:35Z","lastTransitionTime":"2025-12-03T16:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.831021 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.831065 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.831076 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.831092 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.831104 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:35Z","lastTransitionTime":"2025-12-03T16:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.933493 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.933917 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.933932 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.933949 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:35 crc kubenswrapper[4998]: I1203 16:04:35.933960 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:35Z","lastTransitionTime":"2025-12-03T16:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.037064 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.037099 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.037108 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.037121 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.037130 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:36Z","lastTransitionTime":"2025-12-03T16:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.048994 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m2lh2_4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9/ovnkube-controller/1.log" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.052246 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" event={"ID":"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9","Type":"ContainerStarted","Data":"05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958"} Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.053249 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.068771 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:36Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.081267 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:36Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.097246 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k8ptd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2914eb52-522e-4789-a988-b15875755144\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k8ptd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:36Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.111025 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ece18388-e2af-4428-bd37-0835a7dfebd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff6d272ee68c57986031e2d525eced78a6c7316a8980d1bae00590e14ce4fcdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b5dff655685383dcbeeb2a97f745bc3708c15cdfcf113ec1e2b230daf70bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235459dff78aa6c80163b68bdb48c604268b9f8bb340326a52de53d44bcb98bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f6817f5eb33e35e106cb224f0fd8a31f3ff73d3a12621774680bbec5d24a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92f6817f5eb33e35e106cb224f0fd8a31f3ff73d3a12621774680bbec5d24a2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:36Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.131477 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:36Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.140028 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.140099 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.140120 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.140153 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.140173 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:36Z","lastTransitionTime":"2025-12-03T16:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.154125 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:36Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.174163 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d856098-4acb-4d57-8a27-de806955228c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53614cedb63f775d2b563e4405055cd6702c3f35980bc092ac07a3c9ff69aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c911081098fed822c164c7e59c1518ec11cf1c612e3ac4c84af8e694498c915d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9cdcf95d23503cce6d3fbc6de109783933aaaec4eefccd6f1834e234ed41f6e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:36Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.197717 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:36Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.219845 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e48502ef54b1ba19755467c5f08d3596567a1f41f37ed5fdf117db7e617b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:36Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.232888 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65cf533-fd60-47a3-aab2-b7377bab9f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44d20df9678d9388f9611bf182cf962781bbcb668bccef2fba40783ca6cd7cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93a183a13918fa4a6c42b6d0feed52f46b3eac00d85977936772b4cd7e4cb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47k9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:36Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.242917 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.242957 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.242968 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.242983 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.242996 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:36Z","lastTransitionTime":"2025-12-03T16:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.254488 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:36Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.271939 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:36Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.284845 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:36Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.302918 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"message\\\":\\\"ler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 16:04:11.823033 6464 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:04:11.823064 6464 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:04:11.823080 6464 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:04:11.823178 6464 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:11.823309 6464 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823440 6464 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823488 6464 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823628 6464 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:04:11.823784 6464 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:11.824254 6464 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:36Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.320510 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:36Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.333263 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:36Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.345130 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.345179 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.345189 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.345207 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.345218 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:36Z","lastTransitionTime":"2025-12-03T16:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.345941 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:36Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.360209 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:36Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.447945 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.447981 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.447990 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.448004 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.448012 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:36Z","lastTransitionTime":"2025-12-03T16:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.551427 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.551488 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.551509 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.551536 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.551555 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:36Z","lastTransitionTime":"2025-12-03T16:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.653881 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.653924 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.653934 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.653950 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.653960 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:36Z","lastTransitionTime":"2025-12-03T16:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.677125 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:36 crc kubenswrapper[4998]: E1203 16:04:36.677222 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.677386 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:36 crc kubenswrapper[4998]: E1203 16:04:36.677449 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.677545 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:36 crc kubenswrapper[4998]: E1203 16:04:36.677688 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.755840 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.756155 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.756253 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.756343 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.756435 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:36Z","lastTransitionTime":"2025-12-03T16:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.859021 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.859073 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.859086 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.859107 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.859121 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:36Z","lastTransitionTime":"2025-12-03T16:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.962079 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.962163 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.962190 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.962227 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:36 crc kubenswrapper[4998]: I1203 16:04:36.962257 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:36Z","lastTransitionTime":"2025-12-03T16:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.061669 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m2lh2_4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9/ovnkube-controller/2.log" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.062582 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m2lh2_4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9/ovnkube-controller/1.log" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.064404 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.064483 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.064511 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.064561 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.064586 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:37Z","lastTransitionTime":"2025-12-03T16:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.067211 4998 generic.go:334] "Generic (PLEG): container finished" podID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerID="05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958" exitCode=1 Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.067285 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" event={"ID":"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9","Type":"ContainerDied","Data":"05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958"} Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.067359 4998 scope.go:117] "RemoveContainer" containerID="609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.069351 4998 scope.go:117] "RemoveContainer" containerID="05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958" Dec 03 16:04:37 crc kubenswrapper[4998]: E1203 16:04:37.069593 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-m2lh2_openshift-ovn-kubernetes(4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.086106 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ece18388-e2af-4428-bd37-0835a7dfebd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff6d272ee68c57986031e2d525eced78a6c7316a8980d1bae00590e14ce4fcdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b5dff655685383dcbeeb2a97f745bc3708c15cdfcf113ec1e2b230daf70bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235459dff78aa6c80163b68bdb48c604268b9f8bb340326a52de53d44bcb98bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f6817f5eb33e35e106cb224f0fd8a31f3ff73d3a12621774680bbec5d24a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92f6817f5eb33e35e106cb224f0fd8a31f3ff73d3a12621774680bbec5d24a2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:37Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.109446 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:37Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.123260 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:37Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.137525 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:37Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.153409 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:37Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.165729 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k8ptd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2914eb52-522e-4789-a988-b15875755144\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k8ptd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:37Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.168126 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.168180 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.168199 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.168227 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.168245 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:37Z","lastTransitionTime":"2025-12-03T16:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.183841 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d856098-4acb-4d57-8a27-de806955228c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53614cedb63f775d2b563e4405055cd6702c3f35980bc092ac07a3c9ff69aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c911081098fed822c164c7e59c1518ec11cf1c612e3ac4c84af8e694498c915d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9cdcf95d23503cce6d3fbc6de109783933aaaec4eefccd6f1834e234ed41f6e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:37Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.197773 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:37Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.217039 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e48502ef54b1ba19755467c5f08d3596567a1f41f37ed5fdf117db7e617b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:37Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.232252 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65cf533-fd60-47a3-aab2-b7377bab9f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44d20df9678d9388f9611bf182cf962781bbcb668bccef2fba40783ca6cd7cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93a183a13918fa4a6c42b6d0feed52f46b3eac00d85977936772b4cd7e4cb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47k9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:37Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.250726 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:37Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.267275 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:37Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.271801 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.271881 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.271895 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.271916 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.271931 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:37Z","lastTransitionTime":"2025-12-03T16:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.300474 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:37Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.324720 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:37Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.345450 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:37Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.367046 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:37Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.374994 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.375044 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.375062 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.375086 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.375103 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:37Z","lastTransitionTime":"2025-12-03T16:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.383502 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:37Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.405919 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://609b9b553fb765d73b578555d483aca17ada9c90b64e46a8585d94c74ad1db95\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"message\\\":\\\"ler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 16:04:11.823033 6464 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:04:11.823064 6464 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:04:11.823080 6464 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:04:11.823178 6464 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:11.823309 6464 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823440 6464 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823488 6464 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:11.823628 6464 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:04:11.823784 6464 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:11.824254 6464 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:04:36Z\\\",\\\"message\\\":\\\"dler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 16:04:36.686345 6741 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:04:36.686365 6741 factory.go:656] Stopping watch factory\\\\nI1203 16:04:36.686363 6741 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16:04:36.686373 6741 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:04:36.686373 6741 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:36.686403 6741 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:04:36.686513 6741 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:36.686730 6741 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:36.686815 6741 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:36.686826 6741 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:04:36.686863 6741 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:37Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.478395 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.478455 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.478476 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.478506 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.478528 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:37Z","lastTransitionTime":"2025-12-03T16:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.582117 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.582183 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.582207 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.582238 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.582260 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:37Z","lastTransitionTime":"2025-12-03T16:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.677617 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:37 crc kubenswrapper[4998]: E1203 16:04:37.677896 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.685159 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.685269 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.685296 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.685331 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.685366 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:37Z","lastTransitionTime":"2025-12-03T16:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.788453 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.788517 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.788530 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.788549 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.788563 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:37Z","lastTransitionTime":"2025-12-03T16:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.891110 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.891170 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.891190 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.891216 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.891234 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:37Z","lastTransitionTime":"2025-12-03T16:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.993628 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.993674 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.993685 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.993702 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:37 crc kubenswrapper[4998]: I1203 16:04:37.993714 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:37Z","lastTransitionTime":"2025-12-03T16:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.072587 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m2lh2_4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9/ovnkube-controller/2.log" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.076308 4998 scope.go:117] "RemoveContainer" containerID="05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958" Dec 03 16:04:38 crc kubenswrapper[4998]: E1203 16:04:38.076474 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-m2lh2_openshift-ovn-kubernetes(4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.095493 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.095528 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.095539 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.095555 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.095565 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:38Z","lastTransitionTime":"2025-12-03T16:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.104261 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:38Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.117439 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:38Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.128893 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:38Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.140911 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:38Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.155160 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:38Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.182254 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:04:36Z\\\",\\\"message\\\":\\\"dler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 16:04:36.686345 6741 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:04:36.686365 6741 factory.go:656] Stopping watch factory\\\\nI1203 16:04:36.686363 6741 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16:04:36.686373 6741 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:04:36.686373 6741 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:36.686403 6741 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:04:36.686513 6741 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:36.686730 6741 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:36.686815 6741 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:36.686826 6741 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:04:36.686863 6741 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-m2lh2_openshift-ovn-kubernetes(4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:38Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.196692 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ece18388-e2af-4428-bd37-0835a7dfebd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff6d272ee68c57986031e2d525eced78a6c7316a8980d1bae00590e14ce4fcdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b5dff655685383dcbeeb2a97f745bc3708c15cdfcf113ec1e2b230daf70bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235459dff78aa6c80163b68bdb48c604268b9f8bb340326a52de53d44bcb98bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f6817f5eb33e35e106cb224f0fd8a31f3ff73d3a12621774680bbec5d24a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92f6817f5eb33e35e106cb224f0fd8a31f3ff73d3a12621774680bbec5d24a2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:38Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.198510 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.198559 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.198577 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.198596 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.198609 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:38Z","lastTransitionTime":"2025-12-03T16:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.212642 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:38Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.229030 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:38Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.241503 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:38Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.253015 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:38Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.268822 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k8ptd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2914eb52-522e-4789-a988-b15875755144\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k8ptd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:38Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.287391 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d856098-4acb-4d57-8a27-de806955228c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53614cedb63f775d2b563e4405055cd6702c3f35980bc092ac07a3c9ff69aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c911081098fed822c164c7e59c1518ec11cf1c612e3ac4c84af8e694498c915d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9cdcf95d23503cce6d3fbc6de109783933aaaec4eefccd6f1834e234ed41f6e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:38Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.302157 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.302222 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.302244 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.302276 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.302299 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:38Z","lastTransitionTime":"2025-12-03T16:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.307022 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:38Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.325688 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e48502ef54b1ba19755467c5f08d3596567a1f41f37ed5fdf117db7e617b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:38Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.340286 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65cf533-fd60-47a3-aab2-b7377bab9f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44d20df9678d9388f9611bf182cf962781bbcb668bccef2fba40783ca6cd7cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93a183a13918fa4a6c42b6d0feed52f46b3eac00d85977936772b4cd7e4cb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47k9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:38Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.354387 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:38Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.368372 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:38Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.405365 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.405459 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.405485 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.405518 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.405545 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:38Z","lastTransitionTime":"2025-12-03T16:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.508796 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.508853 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.508865 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.508883 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.508900 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:38Z","lastTransitionTime":"2025-12-03T16:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.612109 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.612180 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.612204 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.612236 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.612258 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:38Z","lastTransitionTime":"2025-12-03T16:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.677178 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.677230 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.677205 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:38 crc kubenswrapper[4998]: E1203 16:04:38.677353 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:38 crc kubenswrapper[4998]: E1203 16:04:38.677569 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:38 crc kubenswrapper[4998]: E1203 16:04:38.677830 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.714568 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.714619 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.714637 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.714660 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.714677 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:38Z","lastTransitionTime":"2025-12-03T16:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.818540 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.818601 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.818613 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.818638 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.818655 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:38Z","lastTransitionTime":"2025-12-03T16:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.922347 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.922402 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.922417 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.922441 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:38 crc kubenswrapper[4998]: I1203 16:04:38.922459 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:38Z","lastTransitionTime":"2025-12-03T16:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.025031 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.025105 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.025120 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.025144 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.025164 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:39Z","lastTransitionTime":"2025-12-03T16:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.128750 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.128804 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.128812 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.128828 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.128839 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:39Z","lastTransitionTime":"2025-12-03T16:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.232183 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.232235 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.232253 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.232276 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.232293 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:39Z","lastTransitionTime":"2025-12-03T16:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.335421 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.335530 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.335549 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.335570 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.335586 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:39Z","lastTransitionTime":"2025-12-03T16:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.438792 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.438833 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.438844 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.438861 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.438875 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:39Z","lastTransitionTime":"2025-12-03T16:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.542155 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.542221 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.542237 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.542262 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.542279 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:39Z","lastTransitionTime":"2025-12-03T16:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.646052 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.646132 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.646155 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.646188 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.646212 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:39Z","lastTransitionTime":"2025-12-03T16:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.677577 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:39 crc kubenswrapper[4998]: E1203 16:04:39.677770 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.750012 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.750062 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.750074 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.750091 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.750103 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:39Z","lastTransitionTime":"2025-12-03T16:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.852872 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.852933 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.852951 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.852976 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.852993 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:39Z","lastTransitionTime":"2025-12-03T16:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.956088 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.956196 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.956221 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.956256 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:39 crc kubenswrapper[4998]: I1203 16:04:39.956283 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:39Z","lastTransitionTime":"2025-12-03T16:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.059991 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.060086 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.060111 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.060143 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.060167 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:40Z","lastTransitionTime":"2025-12-03T16:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.163124 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.163172 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.163182 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.163199 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.163210 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:40Z","lastTransitionTime":"2025-12-03T16:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.266445 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.266505 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.266521 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.266544 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.266561 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:40Z","lastTransitionTime":"2025-12-03T16:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.370559 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.370625 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.370642 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.370669 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.370686 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:40Z","lastTransitionTime":"2025-12-03T16:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.473708 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.473795 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.473816 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.473841 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.473858 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:40Z","lastTransitionTime":"2025-12-03T16:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.577250 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.577308 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.577322 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.577347 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.577366 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:40Z","lastTransitionTime":"2025-12-03T16:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.676735 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.676853 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.676809 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:40 crc kubenswrapper[4998]: E1203 16:04:40.676998 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:40 crc kubenswrapper[4998]: E1203 16:04:40.677147 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:40 crc kubenswrapper[4998]: E1203 16:04:40.677317 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.679698 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.679741 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.679772 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.679792 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.679803 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:40Z","lastTransitionTime":"2025-12-03T16:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.783397 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.783464 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.783489 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.783520 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.783547 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:40Z","lastTransitionTime":"2025-12-03T16:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.887116 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.887183 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.887207 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.887238 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.887262 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:40Z","lastTransitionTime":"2025-12-03T16:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.990489 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.990570 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.990596 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.990621 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:40 crc kubenswrapper[4998]: I1203 16:04:40.990643 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:40Z","lastTransitionTime":"2025-12-03T16:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.093158 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.093207 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.093218 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.093236 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.093248 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:41Z","lastTransitionTime":"2025-12-03T16:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.195134 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.195193 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.195205 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.195222 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.195233 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:41Z","lastTransitionTime":"2025-12-03T16:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:41 crc kubenswrapper[4998]: E1203 16:04:41.209705 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:41Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.214456 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.214497 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.214509 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.214526 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.214538 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:41Z","lastTransitionTime":"2025-12-03T16:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:41 crc kubenswrapper[4998]: E1203 16:04:41.232775 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:41Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.237340 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.237478 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.237594 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.237707 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.237797 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:41Z","lastTransitionTime":"2025-12-03T16:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:41 crc kubenswrapper[4998]: E1203 16:04:41.258090 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:41Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.262437 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.262515 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.262541 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.262571 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.262595 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:41Z","lastTransitionTime":"2025-12-03T16:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:41 crc kubenswrapper[4998]: E1203 16:04:41.278506 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:41Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.283740 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.283802 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.283821 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.283841 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.283856 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:41Z","lastTransitionTime":"2025-12-03T16:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:41 crc kubenswrapper[4998]: E1203 16:04:41.299952 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:41Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:41 crc kubenswrapper[4998]: E1203 16:04:41.300351 4998 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.302339 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.302457 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.302537 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.302630 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.302704 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:41Z","lastTransitionTime":"2025-12-03T16:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.405800 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.405845 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.405860 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.405879 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.405891 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:41Z","lastTransitionTime":"2025-12-03T16:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.508259 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.508313 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.508332 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.508355 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.508372 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:41Z","lastTransitionTime":"2025-12-03T16:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.611485 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.611534 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.611552 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.611575 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.611594 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:41Z","lastTransitionTime":"2025-12-03T16:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.676947 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:41 crc kubenswrapper[4998]: E1203 16:04:41.677201 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.696050 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65cf533-fd60-47a3-aab2-b7377bab9f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44d20df9678d9388f9611bf182cf962781bbcb668bccef2fba40783ca6cd7cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93a183a13918fa4a6c42b6d0feed52f46b3eac00d85977936772b4cd7e4cb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47k9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:41Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.715266 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.715324 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.715342 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.715366 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.715387 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:41Z","lastTransitionTime":"2025-12-03T16:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.717366 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d856098-4acb-4d57-8a27-de806955228c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53614cedb63f775d2b563e4405055cd6702c3f35980bc092ac07a3c9ff69aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c911081098fed822c164c7e59c1518ec11cf1c612e3ac4c84af8e694498c915d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9cdcf95d23503cce6d3fbc6de109783933aaaec4eefccd6f1834e234ed41f6e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:41Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.734410 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:41Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.763411 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e48502ef54b1ba19755467c5f08d3596567a1f41f37ed5fdf117db7e617b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:41Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.785171 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:41Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.806483 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:41Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.817636 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.817689 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.817704 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.817723 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.817736 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:41Z","lastTransitionTime":"2025-12-03T16:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.826595 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:41Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.841653 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:41Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.861640 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:04:36Z\\\",\\\"message\\\":\\\"dler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 16:04:36.686345 6741 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:04:36.686365 6741 factory.go:656] Stopping watch factory\\\\nI1203 16:04:36.686363 6741 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16:04:36.686373 6741 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:04:36.686373 6741 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:36.686403 6741 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:04:36.686513 6741 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:36.686730 6741 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:36.686815 6741 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:36.686826 6741 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:04:36.686863 6741 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-m2lh2_openshift-ovn-kubernetes(4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:41Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.889651 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:41Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.908533 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:41Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.921022 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.921067 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.921081 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.921099 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.921112 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:41Z","lastTransitionTime":"2025-12-03T16:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.925513 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:41Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.940808 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:41Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.954449 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:41Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.968223 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:41Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:41 crc kubenswrapper[4998]: I1203 16:04:41.984421 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k8ptd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2914eb52-522e-4789-a988-b15875755144\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k8ptd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:41Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.023347 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.023390 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.023403 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.023423 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.023437 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:42Z","lastTransitionTime":"2025-12-03T16:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.043888 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ece18388-e2af-4428-bd37-0835a7dfebd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff6d272ee68c57986031e2d525eced78a6c7316a8980d1bae00590e14ce4fcdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b5dff655685383dcbeeb2a97f745bc3708c15cdfcf113ec1e2b230daf70bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235459dff78aa6c80163b68bdb48c604268b9f8bb340326a52de53d44bcb98bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f6817f5eb33e35e106cb224f0fd8a31f3ff73d3a12621774680bbec5d24a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92f6817f5eb33e35e106cb224f0fd8a31f3ff73d3a12621774680bbec5d24a2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:42Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.060224 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:42Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.125415 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.125490 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.125504 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.125526 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.125541 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:42Z","lastTransitionTime":"2025-12-03T16:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.228793 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.228845 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.229046 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.229066 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.229084 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:42Z","lastTransitionTime":"2025-12-03T16:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.331833 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.331888 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.331906 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.331928 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.331944 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:42Z","lastTransitionTime":"2025-12-03T16:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.434021 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.434092 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.434112 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.434139 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.434157 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:42Z","lastTransitionTime":"2025-12-03T16:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.536337 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.536393 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.536408 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.536428 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.536441 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:42Z","lastTransitionTime":"2025-12-03T16:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.639009 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.639044 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.639053 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.639065 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.639076 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:42Z","lastTransitionTime":"2025-12-03T16:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.677065 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.677084 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.677111 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:42 crc kubenswrapper[4998]: E1203 16:04:42.677177 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:42 crc kubenswrapper[4998]: E1203 16:04:42.677293 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:42 crc kubenswrapper[4998]: E1203 16:04:42.677382 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.742100 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.742133 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.742142 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.742155 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.742164 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:42Z","lastTransitionTime":"2025-12-03T16:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.844919 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.844964 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.844975 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.844991 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.845003 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:42Z","lastTransitionTime":"2025-12-03T16:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.947394 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.947437 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.947449 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.947467 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:42 crc kubenswrapper[4998]: I1203 16:04:42.947478 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:42Z","lastTransitionTime":"2025-12-03T16:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.050062 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.050115 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.050127 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.050147 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.050160 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:43Z","lastTransitionTime":"2025-12-03T16:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.153322 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.153352 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.153360 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.153375 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.153385 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:43Z","lastTransitionTime":"2025-12-03T16:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.255531 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.255579 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.255591 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.255608 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.255619 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:43Z","lastTransitionTime":"2025-12-03T16:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.358788 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.358832 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.358843 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.358860 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.358872 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:43Z","lastTransitionTime":"2025-12-03T16:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.462120 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.462187 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.462205 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.462230 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.462246 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:43Z","lastTransitionTime":"2025-12-03T16:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.566580 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.566634 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.566653 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.566679 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.566696 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:43Z","lastTransitionTime":"2025-12-03T16:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.670295 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.670468 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.670501 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.670536 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.670554 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:43Z","lastTransitionTime":"2025-12-03T16:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.677068 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:43 crc kubenswrapper[4998]: E1203 16:04:43.677253 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.775243 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.775322 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.775342 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.775364 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.775413 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:43Z","lastTransitionTime":"2025-12-03T16:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.879481 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.879545 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.879563 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.879588 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.879605 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:43Z","lastTransitionTime":"2025-12-03T16:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.982440 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.982505 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.982523 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.982547 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:43 crc kubenswrapper[4998]: I1203 16:04:43.982567 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:43Z","lastTransitionTime":"2025-12-03T16:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.086047 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.086095 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.086110 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.086133 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.086150 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:44Z","lastTransitionTime":"2025-12-03T16:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.188977 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.189027 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.189045 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.189069 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.189087 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:44Z","lastTransitionTime":"2025-12-03T16:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.291853 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.291901 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.291918 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.291941 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.291959 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:44Z","lastTransitionTime":"2025-12-03T16:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.394686 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.394787 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.394806 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.394832 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.394851 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:44Z","lastTransitionTime":"2025-12-03T16:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.497675 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.497737 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.497788 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.497813 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.497829 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:44Z","lastTransitionTime":"2025-12-03T16:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.600268 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.600331 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.600359 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.600390 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.600411 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:44Z","lastTransitionTime":"2025-12-03T16:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.676658 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.676841 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.677017 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:44 crc kubenswrapper[4998]: E1203 16:04:44.677095 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:44 crc kubenswrapper[4998]: E1203 16:04:44.676993 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:44 crc kubenswrapper[4998]: E1203 16:04:44.677244 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.702890 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.702964 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.702989 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.703018 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.703038 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:44Z","lastTransitionTime":"2025-12-03T16:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.805482 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.805528 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.805541 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.805558 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.805569 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:44Z","lastTransitionTime":"2025-12-03T16:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.908418 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.908477 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.908496 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.908525 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:44 crc kubenswrapper[4998]: I1203 16:04:44.908543 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:44Z","lastTransitionTime":"2025-12-03T16:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.011378 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.011417 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.011426 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.011441 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.011454 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:45Z","lastTransitionTime":"2025-12-03T16:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.114599 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.114644 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.114659 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.114676 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.114688 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:45Z","lastTransitionTime":"2025-12-03T16:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.217249 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.217286 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.217296 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.217309 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.217318 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:45Z","lastTransitionTime":"2025-12-03T16:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.322953 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.323030 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.323053 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.323084 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.323106 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:45Z","lastTransitionTime":"2025-12-03T16:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.426328 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.426373 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.426384 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.426401 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.426413 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:45Z","lastTransitionTime":"2025-12-03T16:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.531580 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.531618 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.531628 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.531646 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.531657 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:45Z","lastTransitionTime":"2025-12-03T16:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.633526 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.633579 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.633590 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.633607 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.633618 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:45Z","lastTransitionTime":"2025-12-03T16:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.677411 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:45 crc kubenswrapper[4998]: E1203 16:04:45.677562 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.736387 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.736454 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.736471 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.736495 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.736517 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:45Z","lastTransitionTime":"2025-12-03T16:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.838945 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.838977 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.838987 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.839000 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.839009 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:45Z","lastTransitionTime":"2025-12-03T16:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.941160 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.941201 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.941210 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.941223 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:45 crc kubenswrapper[4998]: I1203 16:04:45.941233 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:45Z","lastTransitionTime":"2025-12-03T16:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.043463 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.043506 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.043517 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.043534 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.043546 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:46Z","lastTransitionTime":"2025-12-03T16:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.145322 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.145366 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.145378 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.145394 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.145405 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:46Z","lastTransitionTime":"2025-12-03T16:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.191340 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2914eb52-522e-4789-a988-b15875755144-metrics-certs\") pod \"network-metrics-daemon-k8ptd\" (UID: \"2914eb52-522e-4789-a988-b15875755144\") " pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:46 crc kubenswrapper[4998]: E1203 16:04:46.191507 4998 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:04:46 crc kubenswrapper[4998]: E1203 16:04:46.191577 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2914eb52-522e-4789-a988-b15875755144-metrics-certs podName:2914eb52-522e-4789-a988-b15875755144 nodeName:}" failed. No retries permitted until 2025-12-03 16:05:18.191558278 +0000 UTC m=+96.803258501 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2914eb52-522e-4789-a988-b15875755144-metrics-certs") pod "network-metrics-daemon-k8ptd" (UID: "2914eb52-522e-4789-a988-b15875755144") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.247922 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.247968 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.247980 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.247996 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.248008 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:46Z","lastTransitionTime":"2025-12-03T16:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.349916 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.349960 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.349971 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.349987 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.349998 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:46Z","lastTransitionTime":"2025-12-03T16:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.452331 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.452390 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.452407 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.452430 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.452449 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:46Z","lastTransitionTime":"2025-12-03T16:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.554963 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.555012 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.555026 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.555043 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.555055 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:46Z","lastTransitionTime":"2025-12-03T16:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.657639 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.657679 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.657690 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.657707 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.657720 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:46Z","lastTransitionTime":"2025-12-03T16:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.677558 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:46 crc kubenswrapper[4998]: E1203 16:04:46.677672 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.677766 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.677779 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:46 crc kubenswrapper[4998]: E1203 16:04:46.677906 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:46 crc kubenswrapper[4998]: E1203 16:04:46.677932 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.759619 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.759663 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.759680 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.759700 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.759716 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:46Z","lastTransitionTime":"2025-12-03T16:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.862131 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.862173 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.862183 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.862199 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.862209 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:46Z","lastTransitionTime":"2025-12-03T16:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.964882 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.964940 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.964957 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.964981 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:46 crc kubenswrapper[4998]: I1203 16:04:46.964998 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:46Z","lastTransitionTime":"2025-12-03T16:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.068036 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.068093 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.068109 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.068133 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.068151 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:47Z","lastTransitionTime":"2025-12-03T16:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.170718 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.170781 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.170793 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.170815 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.170827 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:47Z","lastTransitionTime":"2025-12-03T16:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.273352 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.273392 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.273401 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.273417 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.273426 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:47Z","lastTransitionTime":"2025-12-03T16:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.376771 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.376817 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.376825 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.376841 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.376850 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:47Z","lastTransitionTime":"2025-12-03T16:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.479111 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.479161 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.479170 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.479187 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.479199 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:47Z","lastTransitionTime":"2025-12-03T16:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.581405 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.581464 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.581485 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.581510 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.581528 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:47Z","lastTransitionTime":"2025-12-03T16:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.676747 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:47 crc kubenswrapper[4998]: E1203 16:04:47.676930 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.684353 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.684391 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.684401 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.684417 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.684430 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:47Z","lastTransitionTime":"2025-12-03T16:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.787330 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.787388 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.787407 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.787431 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.787449 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:47Z","lastTransitionTime":"2025-12-03T16:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.890039 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.890081 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.890098 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.890121 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.890137 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:47Z","lastTransitionTime":"2025-12-03T16:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.993242 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.993304 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.993321 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.993344 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:47 crc kubenswrapper[4998]: I1203 16:04:47.993361 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:47Z","lastTransitionTime":"2025-12-03T16:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.095478 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.095512 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.095523 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.095536 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.095547 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:48Z","lastTransitionTime":"2025-12-03T16:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.115780 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fcdxh_bd9d66fb-a400-4810-aa7a-c81c9c24bd11/kube-multus/0.log" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.115850 4998 generic.go:334] "Generic (PLEG): container finished" podID="bd9d66fb-a400-4810-aa7a-c81c9c24bd11" containerID="d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d" exitCode=1 Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.115898 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fcdxh" event={"ID":"bd9d66fb-a400-4810-aa7a-c81c9c24bd11","Type":"ContainerDied","Data":"d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d"} Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.117042 4998 scope.go:117] "RemoveContainer" containerID="d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.133676 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:48Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.151056 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:48Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.162136 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:48Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.174235 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:48Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.190716 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k8ptd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2914eb52-522e-4789-a988-b15875755144\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k8ptd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:48Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.197842 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.197988 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.198150 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.198241 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.198455 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:48Z","lastTransitionTime":"2025-12-03T16:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.204126 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ece18388-e2af-4428-bd37-0835a7dfebd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff6d272ee68c57986031e2d525eced78a6c7316a8980d1bae00590e14ce4fcdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b5dff655685383dcbeeb2a97f745bc3708c15cdfcf113ec1e2b230daf70bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235459dff78aa6c80163b68bdb48c604268b9f8bb340326a52de53d44bcb98bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f6817f5eb33e35e106cb224f0fd8a31f3ff73d3a12621774680bbec5d24a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92f6817f5eb33e35e106cb224f0fd8a31f3ff73d3a12621774680bbec5d24a2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:48Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.220194 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e48502ef54b1ba19755467c5f08d3596567a1f41f37ed5fdf117db7e617b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:48Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.231287 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65cf533-fd60-47a3-aab2-b7377bab9f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44d20df9678d9388f9611bf182cf962781bbcb668bccef2fba40783ca6cd7cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93a183a13918fa4a6c42b6d0feed52f46b3eac00d85977936772b4cd7e4cb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47k9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:48Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.242954 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d856098-4acb-4d57-8a27-de806955228c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53614cedb63f775d2b563e4405055cd6702c3f35980bc092ac07a3c9ff69aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c911081098fed822c164c7e59c1518ec11cf1c612e3ac4c84af8e694498c915d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9cdcf95d23503cce6d3fbc6de109783933aaaec4eefccd6f1834e234ed41f6e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:48Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.260928 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:48Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.273250 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:48Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.288527 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:04:47Z\\\",\\\"message\\\":\\\"2025-12-03T16:04:01+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a1fbe999-84ce-42e2-80f8-36647b645314\\\\n2025-12-03T16:04:01+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a1fbe999-84ce-42e2-80f8-36647b645314 to /host/opt/cni/bin/\\\\n2025-12-03T16:04:01Z [verbose] multus-daemon started\\\\n2025-12-03T16:04:01Z [verbose] Readiness Indicator file check\\\\n2025-12-03T16:04:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:48Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.301459 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.301493 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.301506 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.301523 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.301537 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:48Z","lastTransitionTime":"2025-12-03T16:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.309269 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:48Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.321800 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:48Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.332684 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:48Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.352329 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:04:36Z\\\",\\\"message\\\":\\\"dler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 16:04:36.686345 6741 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:04:36.686365 6741 factory.go:656] Stopping watch factory\\\\nI1203 16:04:36.686363 6741 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16:04:36.686373 6741 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:04:36.686373 6741 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:36.686403 6741 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:04:36.686513 6741 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:36.686730 6741 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:36.686815 6741 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:36.686826 6741 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:04:36.686863 6741 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-m2lh2_openshift-ovn-kubernetes(4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:48Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.374149 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:48Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.386554 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:48Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.404384 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.404433 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.404444 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.404461 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.404471 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:48Z","lastTransitionTime":"2025-12-03T16:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.507197 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.507249 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.507268 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.507290 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.507307 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:48Z","lastTransitionTime":"2025-12-03T16:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.611293 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.611348 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.611366 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.611389 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.611406 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:48Z","lastTransitionTime":"2025-12-03T16:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.677694 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.677746 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.677803 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:48 crc kubenswrapper[4998]: E1203 16:04:48.678129 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:48 crc kubenswrapper[4998]: E1203 16:04:48.677959 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:48 crc kubenswrapper[4998]: E1203 16:04:48.678305 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.713573 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.713628 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.713646 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.713672 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.713698 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:48Z","lastTransitionTime":"2025-12-03T16:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.816791 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.816849 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.816867 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.816891 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.816909 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:48Z","lastTransitionTime":"2025-12-03T16:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.919103 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.919147 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.919155 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.919170 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:48 crc kubenswrapper[4998]: I1203 16:04:48.919179 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:48Z","lastTransitionTime":"2025-12-03T16:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.021354 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.021404 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.021416 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.021437 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.021448 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:49Z","lastTransitionTime":"2025-12-03T16:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.121616 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fcdxh_bd9d66fb-a400-4810-aa7a-c81c9c24bd11/kube-multus/0.log" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.121663 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fcdxh" event={"ID":"bd9d66fb-a400-4810-aa7a-c81c9c24bd11","Type":"ContainerStarted","Data":"1341e3406e28b4e4309aaa18fba25c82ff489cc9cb44f232c4eb0f9a3762205c"} Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.123556 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.123581 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.123588 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.123599 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.123609 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:49Z","lastTransitionTime":"2025-12-03T16:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.134464 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:49Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.150175 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:49Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.174080 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:04:36Z\\\",\\\"message\\\":\\\"dler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 16:04:36.686345 6741 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:04:36.686365 6741 factory.go:656] Stopping watch factory\\\\nI1203 16:04:36.686363 6741 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16:04:36.686373 6741 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:04:36.686373 6741 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:36.686403 6741 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:04:36.686513 6741 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:36.686730 6741 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:36.686815 6741 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:36.686826 6741 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:04:36.686863 6741 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-m2lh2_openshift-ovn-kubernetes(4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:49Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.200929 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:49Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.213920 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:49Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.226231 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.226264 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.226276 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.226294 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.226306 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:49Z","lastTransitionTime":"2025-12-03T16:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.227849 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:49Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.245353 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:49Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.256112 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:49Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.266677 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:49Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.278899 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k8ptd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2914eb52-522e-4789-a988-b15875755144\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k8ptd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:49Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.291011 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ece18388-e2af-4428-bd37-0835a7dfebd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff6d272ee68c57986031e2d525eced78a6c7316a8980d1bae00590e14ce4fcdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b5dff655685383dcbeeb2a97f745bc3708c15cdfcf113ec1e2b230daf70bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235459dff78aa6c80163b68bdb48c604268b9f8bb340326a52de53d44bcb98bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f6817f5eb33e35e106cb224f0fd8a31f3ff73d3a12621774680bbec5d24a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92f6817f5eb33e35e106cb224f0fd8a31f3ff73d3a12621774680bbec5d24a2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:49Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.303682 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:49Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.314800 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65cf533-fd60-47a3-aab2-b7377bab9f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44d20df9678d9388f9611bf182cf962781bbcb668bccef2fba40783ca6cd7cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93a183a13918fa4a6c42b6d0feed52f46b3eac00d85977936772b4cd7e4cb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47k9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:49Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.328801 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d856098-4acb-4d57-8a27-de806955228c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53614cedb63f775d2b563e4405055cd6702c3f35980bc092ac07a3c9ff69aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c911081098fed822c164c7e59c1518ec11cf1c612e3ac4c84af8e694498c915d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9cdcf95d23503cce6d3fbc6de109783933aaaec4eefccd6f1834e234ed41f6e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:49Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.328858 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.329040 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.329050 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.329065 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.329073 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:49Z","lastTransitionTime":"2025-12-03T16:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.341169 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:49Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.355705 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e48502ef54b1ba19755467c5f08d3596567a1f41f37ed5fdf117db7e617b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:49Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.366994 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:49Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.377395 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1341e3406e28b4e4309aaa18fba25c82ff489cc9cb44f232c4eb0f9a3762205c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:04:47Z\\\",\\\"message\\\":\\\"2025-12-03T16:04:01+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a1fbe999-84ce-42e2-80f8-36647b645314\\\\n2025-12-03T16:04:01+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a1fbe999-84ce-42e2-80f8-36647b645314 to /host/opt/cni/bin/\\\\n2025-12-03T16:04:01Z [verbose] multus-daemon started\\\\n2025-12-03T16:04:01Z [verbose] Readiness Indicator file check\\\\n2025-12-03T16:04:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:49Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.431930 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.431983 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.431994 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.432010 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.432019 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:49Z","lastTransitionTime":"2025-12-03T16:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.534904 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.534963 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.534979 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.535001 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.535019 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:49Z","lastTransitionTime":"2025-12-03T16:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.637195 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.637311 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.637321 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.637335 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.637345 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:49Z","lastTransitionTime":"2025-12-03T16:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.677195 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:49 crc kubenswrapper[4998]: E1203 16:04:49.677492 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.740165 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.740229 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.740240 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.740255 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.740266 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:49Z","lastTransitionTime":"2025-12-03T16:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.842309 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.842354 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.842365 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.842383 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.842395 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:49Z","lastTransitionTime":"2025-12-03T16:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.944829 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.944887 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.944902 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.944923 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:49 crc kubenswrapper[4998]: I1203 16:04:49.944939 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:49Z","lastTransitionTime":"2025-12-03T16:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.047330 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.047383 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.047406 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.047433 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.047453 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:50Z","lastTransitionTime":"2025-12-03T16:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.150253 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.150313 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.150333 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.150359 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.150376 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:50Z","lastTransitionTime":"2025-12-03T16:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.253078 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.253125 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.253138 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.253157 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.253174 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:50Z","lastTransitionTime":"2025-12-03T16:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.355515 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.355548 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.355557 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.355570 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.355582 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:50Z","lastTransitionTime":"2025-12-03T16:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.459052 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.459108 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.459127 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.459151 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.459167 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:50Z","lastTransitionTime":"2025-12-03T16:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.562021 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.562082 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.562100 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.562127 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.562146 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:50Z","lastTransitionTime":"2025-12-03T16:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.664944 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.664985 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.664996 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.665013 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.665026 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:50Z","lastTransitionTime":"2025-12-03T16:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.677197 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.677217 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:50 crc kubenswrapper[4998]: E1203 16:04:50.677416 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.677550 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:50 crc kubenswrapper[4998]: E1203 16:04:50.677819 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:50 crc kubenswrapper[4998]: E1203 16:04:50.677910 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.693486 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.768628 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.768685 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.768701 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.768725 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.768740 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:50Z","lastTransitionTime":"2025-12-03T16:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.871653 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.871735 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.871796 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.871831 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.871853 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:50Z","lastTransitionTime":"2025-12-03T16:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.974836 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.974895 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.974914 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.974938 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:50 crc kubenswrapper[4998]: I1203 16:04:50.974956 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:50Z","lastTransitionTime":"2025-12-03T16:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.078184 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.078230 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.078238 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.078253 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.078265 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:51Z","lastTransitionTime":"2025-12-03T16:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.180702 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.180779 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.180801 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.180823 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.180835 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:51Z","lastTransitionTime":"2025-12-03T16:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.282685 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.282719 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.282728 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.282744 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.282769 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:51Z","lastTransitionTime":"2025-12-03T16:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.385205 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.385247 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.385261 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.385474 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.385701 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:51Z","lastTransitionTime":"2025-12-03T16:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.489573 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.489636 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.489654 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.489678 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.489696 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:51Z","lastTransitionTime":"2025-12-03T16:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.592883 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.592935 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.592951 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.592970 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.592985 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:51Z","lastTransitionTime":"2025-12-03T16:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.596981 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.597067 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.597093 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.597126 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.597151 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:51Z","lastTransitionTime":"2025-12-03T16:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:51 crc kubenswrapper[4998]: E1203 16:04:51.612909 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.617979 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.618051 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.618077 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.618109 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.618135 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:51Z","lastTransitionTime":"2025-12-03T16:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:51 crc kubenswrapper[4998]: E1203 16:04:51.632420 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.637534 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.637626 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.637644 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.637669 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.637688 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:51Z","lastTransitionTime":"2025-12-03T16:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:51 crc kubenswrapper[4998]: E1203 16:04:51.659034 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.664200 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.664253 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.664266 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.664288 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.664301 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:51Z","lastTransitionTime":"2025-12-03T16:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.677307 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:51 crc kubenswrapper[4998]: E1203 16:04:51.677545 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:04:51 crc kubenswrapper[4998]: E1203 16:04:51.679008 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.684161 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.684195 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.684208 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.684224 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.684236 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:51Z","lastTransitionTime":"2025-12-03T16:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.694168 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:51 crc kubenswrapper[4998]: E1203 16:04:51.697367 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:51 crc kubenswrapper[4998]: E1203 16:04:51.697515 4998 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.699301 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.699334 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.699344 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.699358 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.699367 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:51Z","lastTransitionTime":"2025-12-03T16:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.708883 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1341e3406e28b4e4309aaa18fba25c82ff489cc9cb44f232c4eb0f9a3762205c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:04:47Z\\\",\\\"message\\\":\\\"2025-12-03T16:04:01+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a1fbe999-84ce-42e2-80f8-36647b645314\\\\n2025-12-03T16:04:01+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a1fbe999-84ce-42e2-80f8-36647b645314 to /host/opt/cni/bin/\\\\n2025-12-03T16:04:01Z [verbose] multus-daemon started\\\\n2025-12-03T16:04:01Z [verbose] Readiness Indicator file check\\\\n2025-12-03T16:04:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.729391 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.743912 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.754904 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.770669 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.783427 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.799018 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:04:36Z\\\",\\\"message\\\":\\\"dler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 16:04:36.686345 6741 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:04:36.686365 6741 factory.go:656] Stopping watch factory\\\\nI1203 16:04:36.686363 6741 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16:04:36.686373 6741 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:04:36.686373 6741 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:36.686403 6741 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:04:36.686513 6741 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:36.686730 6741 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:36.686815 6741 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:36.686826 6741 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:04:36.686863 6741 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-m2lh2_openshift-ovn-kubernetes(4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.801310 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.801360 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.801369 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.801399 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.801409 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:51Z","lastTransitionTime":"2025-12-03T16:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.810415 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k8ptd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2914eb52-522e-4789-a988-b15875755144\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k8ptd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.823312 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ece18388-e2af-4428-bd37-0835a7dfebd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff6d272ee68c57986031e2d525eced78a6c7316a8980d1bae00590e14ce4fcdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b5dff655685383dcbeeb2a97f745bc3708c15cdfcf113ec1e2b230daf70bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235459dff78aa6c80163b68bdb48c604268b9f8bb340326a52de53d44bcb98bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f6817f5eb33e35e106cb224f0fd8a31f3ff73d3a12621774680bbec5d24a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92f6817f5eb33e35e106cb224f0fd8a31f3ff73d3a12621774680bbec5d24a2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.833600 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6a74747-520f-492f-8475-0eb91b8c5f97\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9b7ae668ca10ea755cabe43c56545791b51cb31e3733764494bdb9f8daafe67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d25d7c896b5f131195161820371e78cbefd894d5aaf86bbd17a61ab7f617bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d25d7c896b5f131195161820371e78cbefd894d5aaf86bbd17a61ab7f617bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.846526 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.857814 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.868362 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.877775 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.889984 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d856098-4acb-4d57-8a27-de806955228c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53614cedb63f775d2b563e4405055cd6702c3f35980bc092ac07a3c9ff69aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c911081098fed822c164c7e59c1518ec11cf1c612e3ac4c84af8e694498c915d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9cdcf95d23503cce6d3fbc6de109783933aaaec4eefccd6f1834e234ed41f6e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.903356 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.903402 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.903416 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.903434 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.903447 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:51Z","lastTransitionTime":"2025-12-03T16:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.904503 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.916548 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e48502ef54b1ba19755467c5f08d3596567a1f41f37ed5fdf117db7e617b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:51 crc kubenswrapper[4998]: I1203 16:04:51.928801 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65cf533-fd60-47a3-aab2-b7377bab9f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44d20df9678d9388f9611bf182cf962781bbcb668bccef2fba40783ca6cd7cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93a183a13918fa4a6c42b6d0feed52f46b3eac00d85977936772b4cd7e4cb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47k9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:04:51Z is after 2025-08-24T17:21:41Z" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.007013 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.007061 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.007073 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.007091 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.007104 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:52Z","lastTransitionTime":"2025-12-03T16:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.109557 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.109595 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.109604 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.109619 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.109630 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:52Z","lastTransitionTime":"2025-12-03T16:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.212664 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.212734 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.212789 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.212816 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.212834 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:52Z","lastTransitionTime":"2025-12-03T16:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.315198 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.315231 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.315242 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.315257 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.315268 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:52Z","lastTransitionTime":"2025-12-03T16:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.418459 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.418523 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.418548 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.418603 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.418628 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:52Z","lastTransitionTime":"2025-12-03T16:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.521974 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.522337 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.522504 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.522826 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.522898 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:52Z","lastTransitionTime":"2025-12-03T16:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.625117 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.625179 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.625189 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.625205 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.625217 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:52Z","lastTransitionTime":"2025-12-03T16:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.677175 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.677183 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.677252 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:52 crc kubenswrapper[4998]: E1203 16:04:52.677798 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:52 crc kubenswrapper[4998]: E1203 16:04:52.677905 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:52 crc kubenswrapper[4998]: E1203 16:04:52.677974 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.678315 4998 scope.go:117] "RemoveContainer" containerID="05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958" Dec 03 16:04:52 crc kubenswrapper[4998]: E1203 16:04:52.678577 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-m2lh2_openshift-ovn-kubernetes(4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.728429 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.728490 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.728512 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.728537 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.728559 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:52Z","lastTransitionTime":"2025-12-03T16:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.831719 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.831798 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.831811 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.831827 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.831839 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:52Z","lastTransitionTime":"2025-12-03T16:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.934489 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.934552 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.934572 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.934600 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:52 crc kubenswrapper[4998]: I1203 16:04:52.934621 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:52Z","lastTransitionTime":"2025-12-03T16:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.037155 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.037224 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.037246 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.037275 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.037294 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:53Z","lastTransitionTime":"2025-12-03T16:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.138995 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.139034 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.139047 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.139063 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.139073 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:53Z","lastTransitionTime":"2025-12-03T16:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.241319 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.241368 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.241385 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.241403 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.241414 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:53Z","lastTransitionTime":"2025-12-03T16:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.344364 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.344402 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.344413 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.344431 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.344441 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:53Z","lastTransitionTime":"2025-12-03T16:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.447033 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.447090 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.447112 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.447141 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.447164 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:53Z","lastTransitionTime":"2025-12-03T16:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.550449 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.550505 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.550521 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.550546 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.550562 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:53Z","lastTransitionTime":"2025-12-03T16:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.653284 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.653358 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.653377 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.653400 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.653417 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:53Z","lastTransitionTime":"2025-12-03T16:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.677219 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:53 crc kubenswrapper[4998]: E1203 16:04:53.677345 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.755341 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.755389 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.755401 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.755420 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.755432 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:53Z","lastTransitionTime":"2025-12-03T16:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.858005 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.858033 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.858045 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.858062 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.858073 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:53Z","lastTransitionTime":"2025-12-03T16:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.960439 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.960499 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.960517 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.960542 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:53 crc kubenswrapper[4998]: I1203 16:04:53.960558 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:53Z","lastTransitionTime":"2025-12-03T16:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.062995 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.063030 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.063041 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.078635 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.078676 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:54Z","lastTransitionTime":"2025-12-03T16:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.180701 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.180815 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.180833 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.180857 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.180868 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:54Z","lastTransitionTime":"2025-12-03T16:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.283677 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.283719 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.283733 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.283768 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.283782 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:54Z","lastTransitionTime":"2025-12-03T16:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.386382 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.386430 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.386464 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.386482 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.386493 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:54Z","lastTransitionTime":"2025-12-03T16:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.488888 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.488969 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.489016 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.489051 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.489075 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:54Z","lastTransitionTime":"2025-12-03T16:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.591407 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.591457 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.591472 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.591489 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.591500 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:54Z","lastTransitionTime":"2025-12-03T16:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.677633 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.677674 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.677764 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:54 crc kubenswrapper[4998]: E1203 16:04:54.677795 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:54 crc kubenswrapper[4998]: E1203 16:04:54.677877 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:54 crc kubenswrapper[4998]: E1203 16:04:54.677992 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.693471 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.693518 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.693529 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.693547 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.693560 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:54Z","lastTransitionTime":"2025-12-03T16:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.795929 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.795986 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.795998 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.796017 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.796029 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:54Z","lastTransitionTime":"2025-12-03T16:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.898159 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.898206 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.898221 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.898244 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:54 crc kubenswrapper[4998]: I1203 16:04:54.898260 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:54Z","lastTransitionTime":"2025-12-03T16:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.001409 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.001471 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.001483 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.001501 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.001514 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:55Z","lastTransitionTime":"2025-12-03T16:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.103655 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.103699 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.103710 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.103727 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.103737 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:55Z","lastTransitionTime":"2025-12-03T16:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.205636 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.205674 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.205687 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.205704 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.205716 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:55Z","lastTransitionTime":"2025-12-03T16:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.309099 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.309186 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.309209 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.309237 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.309257 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:55Z","lastTransitionTime":"2025-12-03T16:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.412499 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.412560 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.412576 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.412598 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.412616 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:55Z","lastTransitionTime":"2025-12-03T16:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.515059 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.515131 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.515150 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.515177 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.515198 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:55Z","lastTransitionTime":"2025-12-03T16:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.618703 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.618811 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.618840 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.618873 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.618892 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:55Z","lastTransitionTime":"2025-12-03T16:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.677806 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:55 crc kubenswrapper[4998]: E1203 16:04:55.678007 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.721792 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.721852 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.721871 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.721898 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.721917 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:55Z","lastTransitionTime":"2025-12-03T16:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.824727 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.824820 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.824844 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.824873 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.824896 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:55Z","lastTransitionTime":"2025-12-03T16:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.927914 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.928002 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.928030 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.928064 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:55 crc kubenswrapper[4998]: I1203 16:04:55.928085 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:55Z","lastTransitionTime":"2025-12-03T16:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.031425 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.031498 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.031521 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.031550 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.031571 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:56Z","lastTransitionTime":"2025-12-03T16:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.133994 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.134055 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.134068 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.134083 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.134094 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:56Z","lastTransitionTime":"2025-12-03T16:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.236278 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.236340 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.236353 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.236372 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.236383 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:56Z","lastTransitionTime":"2025-12-03T16:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.338918 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.338992 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.339011 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.339049 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.339069 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:56Z","lastTransitionTime":"2025-12-03T16:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.441983 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.442039 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.442053 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.442072 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.442085 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:56Z","lastTransitionTime":"2025-12-03T16:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.545229 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.545289 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.545308 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.545329 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.545345 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:56Z","lastTransitionTime":"2025-12-03T16:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.650937 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.651063 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.651074 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.651092 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.651106 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:56Z","lastTransitionTime":"2025-12-03T16:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.677744 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.677850 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.677979 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:56 crc kubenswrapper[4998]: E1203 16:04:56.678064 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:56 crc kubenswrapper[4998]: E1203 16:04:56.678203 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:56 crc kubenswrapper[4998]: E1203 16:04:56.678237 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.755264 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.755302 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.755317 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.755342 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.755356 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:56Z","lastTransitionTime":"2025-12-03T16:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.859074 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.859123 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.859143 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.859376 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.859398 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:56Z","lastTransitionTime":"2025-12-03T16:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.961994 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.962062 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.962081 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.962109 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:56 crc kubenswrapper[4998]: I1203 16:04:56.962128 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:56Z","lastTransitionTime":"2025-12-03T16:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.065113 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.065183 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.065203 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.065229 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.065246 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:57Z","lastTransitionTime":"2025-12-03T16:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.167629 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.167693 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.167722 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.167746 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.167799 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:57Z","lastTransitionTime":"2025-12-03T16:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.271398 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.271485 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.271552 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.271584 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.271603 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:57Z","lastTransitionTime":"2025-12-03T16:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.374549 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.374586 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.374597 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.374613 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.374625 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:57Z","lastTransitionTime":"2025-12-03T16:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.476960 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.477043 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.477053 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.477070 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.477081 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:57Z","lastTransitionTime":"2025-12-03T16:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.580189 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.580243 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.580260 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.580283 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.580301 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:57Z","lastTransitionTime":"2025-12-03T16:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.677481 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:57 crc kubenswrapper[4998]: E1203 16:04:57.677721 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.683374 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.683441 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.683468 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.683498 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.683521 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:57Z","lastTransitionTime":"2025-12-03T16:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.786114 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.786187 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.786208 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.786235 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.786254 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:57Z","lastTransitionTime":"2025-12-03T16:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.889669 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.889734 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.889788 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.889815 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.889836 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:57Z","lastTransitionTime":"2025-12-03T16:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.993196 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.993269 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.993289 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.993316 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:57 crc kubenswrapper[4998]: I1203 16:04:57.993334 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:57Z","lastTransitionTime":"2025-12-03T16:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.096694 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.096748 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.096796 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.096827 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.096847 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:58Z","lastTransitionTime":"2025-12-03T16:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.200348 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.200417 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.200434 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.200458 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.200476 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:58Z","lastTransitionTime":"2025-12-03T16:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.303449 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.303554 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.303575 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.303603 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.303624 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:58Z","lastTransitionTime":"2025-12-03T16:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.405954 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.406020 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.406039 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.406065 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.406083 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:58Z","lastTransitionTime":"2025-12-03T16:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.508661 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.508697 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.508722 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.508735 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.508745 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:58Z","lastTransitionTime":"2025-12-03T16:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.611344 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.611415 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.611440 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.611469 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.611493 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:58Z","lastTransitionTime":"2025-12-03T16:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.677687 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.677750 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.677679 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:04:58 crc kubenswrapper[4998]: E1203 16:04:58.677949 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:04:58 crc kubenswrapper[4998]: E1203 16:04:58.678077 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:04:58 crc kubenswrapper[4998]: E1203 16:04:58.678171 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.714007 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.714068 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.714088 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.714113 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.714129 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:58Z","lastTransitionTime":"2025-12-03T16:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.817151 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.817215 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.817238 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.817270 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.817294 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:58Z","lastTransitionTime":"2025-12-03T16:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.920087 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.920148 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.920170 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.920196 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:58 crc kubenswrapper[4998]: I1203 16:04:58.920216 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:58Z","lastTransitionTime":"2025-12-03T16:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.022942 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.022981 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.022991 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.023005 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.023013 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:59Z","lastTransitionTime":"2025-12-03T16:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.125736 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.125810 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.125825 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.125845 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.125861 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:59Z","lastTransitionTime":"2025-12-03T16:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.228556 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.228610 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.228627 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.228650 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.228670 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:59Z","lastTransitionTime":"2025-12-03T16:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.331731 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.331788 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.331798 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.331817 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.331833 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:59Z","lastTransitionTime":"2025-12-03T16:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.435538 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.435638 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.435663 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.435690 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.435712 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:59Z","lastTransitionTime":"2025-12-03T16:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.538369 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.538473 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.538542 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.538574 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.538635 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:59Z","lastTransitionTime":"2025-12-03T16:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.642136 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.642181 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.642189 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.642205 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.642235 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:59Z","lastTransitionTime":"2025-12-03T16:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.677093 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:04:59 crc kubenswrapper[4998]: E1203 16:04:59.677256 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.745192 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.745231 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.745416 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.745438 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.745452 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:59Z","lastTransitionTime":"2025-12-03T16:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.848978 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.849046 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.849058 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.849075 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.849086 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:59Z","lastTransitionTime":"2025-12-03T16:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.951601 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.951650 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.951661 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.951679 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:04:59 crc kubenswrapper[4998]: I1203 16:04:59.951690 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:04:59Z","lastTransitionTime":"2025-12-03T16:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.054550 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.054593 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.054605 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.054621 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.054633 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:00Z","lastTransitionTime":"2025-12-03T16:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.158078 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.158142 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.158161 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.158186 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.158205 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:00Z","lastTransitionTime":"2025-12-03T16:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.262975 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.263032 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.263048 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.263076 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.263092 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:00Z","lastTransitionTime":"2025-12-03T16:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.366403 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.366469 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.366490 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.366518 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.366537 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:00Z","lastTransitionTime":"2025-12-03T16:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.469879 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.469956 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.469975 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.470001 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.470020 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:00Z","lastTransitionTime":"2025-12-03T16:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.573627 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.573681 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.573698 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.573721 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.573740 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:00Z","lastTransitionTime":"2025-12-03T16:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.676717 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.676720 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.677290 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.677466 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.677487 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.678568 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.678587 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:00 crc kubenswrapper[4998]: E1203 16:05:00.677468 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.678599 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:00Z","lastTransitionTime":"2025-12-03T16:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:00 crc kubenswrapper[4998]: E1203 16:05:00.678629 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:05:00 crc kubenswrapper[4998]: E1203 16:05:00.678705 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.782246 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.782877 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.782907 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.782927 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.782942 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:00Z","lastTransitionTime":"2025-12-03T16:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.885831 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.885897 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.885914 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.885941 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.885962 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:00Z","lastTransitionTime":"2025-12-03T16:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.988481 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.988527 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.988540 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.988559 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:00 crc kubenswrapper[4998]: I1203 16:05:00.988572 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:00Z","lastTransitionTime":"2025-12-03T16:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.092586 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.092653 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.092671 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.092698 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.092718 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:01Z","lastTransitionTime":"2025-12-03T16:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.196579 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.196644 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.196664 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.196689 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.196707 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:01Z","lastTransitionTime":"2025-12-03T16:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.298414 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.298445 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.298455 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.298468 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.298477 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:01Z","lastTransitionTime":"2025-12-03T16:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.401696 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.401780 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.401794 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.401813 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.401826 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:01Z","lastTransitionTime":"2025-12-03T16:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.505433 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.505864 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.505882 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.505908 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.505926 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:01Z","lastTransitionTime":"2025-12-03T16:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.608494 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.608535 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.608547 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.608563 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.608574 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:01Z","lastTransitionTime":"2025-12-03T16:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.676922 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:05:01 crc kubenswrapper[4998]: E1203 16:05:01.677032 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.699518 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922fdf22837ce952f98745857a0cb5a0cc59b908abf8688cd7ac2e12dd728e3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:05:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.711598 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.711668 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.711686 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.711725 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.711820 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:01Z","lastTransitionTime":"2025-12-03T16:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.718128 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://663ba5f3fea0d13af2fcb4c087ccf3cfebd31298fa2bf4bbd7dd28ecbb1d8225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:05:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.732580 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9srmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5790060f-0fda-43a1-bde3-7332df05db85\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfda9ee3405cf4662c5f018a8698741b192194f6bbb1133305177585c810915d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw4tm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9srmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:05:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.746345 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-74tc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b12292e6-f666-4bdc-8de8-334679d5f89a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4ddbe3fa91422e282abe27ebcc514610a1ff502881fe3bec8150b576e8d9cb7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nftwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-74tc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:05:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.761976 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k8ptd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2914eb52-522e-4789-a988-b15875755144\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zh9r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k8ptd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:05:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.781721 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ece18388-e2af-4428-bd37-0835a7dfebd5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff6d272ee68c57986031e2d525eced78a6c7316a8980d1bae00590e14ce4fcdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://36b5dff655685383dcbeeb2a97f745bc3708c15cdfcf113ec1e2b230daf70bc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235459dff78aa6c80163b68bdb48c604268b9f8bb340326a52de53d44bcb98bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92f6817f5eb33e35e106cb224f0fd8a31f3ff73d3a12621774680bbec5d24a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92f6817f5eb33e35e106cb224f0fd8a31f3ff73d3a12621774680bbec5d24a2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:05:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.797552 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6a74747-520f-492f-8475-0eb91b8c5f97\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c9b7ae668ca10ea755cabe43c56545791b51cb31e3733764494bdb9f8daafe67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d25d7c896b5f131195161820371e78cbefd894d5aaf86bbd17a61ab7f617bb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d25d7c896b5f131195161820371e78cbefd894d5aaf86bbd17a61ab7f617bb2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:05:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.813564 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25e154cc-ca57-431a-bce8-c9e245ec5a1e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://50e48502ef54b1ba19755467c5f08d3596567a1f41f37ed5fdf117db7e617b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80820e94a64861e8ab6ac82ca95d39f6144c8eed6dc04a191015c02ce56c7d0d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d087c6e00727c2d363c7141147629c804ce0559e8cfe45d855ccaea6db2123c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://301d9920d5806bd2c4719429ba76a54f3f873bd99926c060dbf33872bf133b97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3e74345618ace25f04f5addeb587f18888cefa5cb099e95ff91734bcbb4e0c9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7374cc4dcec9e2bca4a151e5572f17e4d0ce8e59460118a1abf9026687646678\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2226ec01f1fccfbdf98a9c2029d2dc050ce6d69b97e32c50d06a5eed250c7300\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8xf9x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qt6sw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:05:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.819015 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.819079 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.819096 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.819120 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.819136 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:01Z","lastTransitionTime":"2025-12-03T16:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.828567 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b65cf533-fd60-47a3-aab2-b7377bab9f9d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44d20df9678d9388f9611bf182cf962781bbcb668bccef2fba40783ca6cd7cbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec93a183a13918fa4a6c42b6d0feed52f46b3eac00d85977936772b4cd7e4cb9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zj529\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47k9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:05:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.842580 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d856098-4acb-4d57-8a27-de806955228c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53614cedb63f775d2b563e4405055cd6702c3f35980bc092ac07a3c9ff69aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c911081098fed822c164c7e59c1518ec11cf1c612e3ac4c84af8e694498c915d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9cdcf95d23503cce6d3fbc6de109783933aaaec4eefccd6f1834e234ed41f6e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:05:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.858119 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17690d1e2312183d3b2beece25a2141eb62847952a95cbf9cd8cda988d3ad559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed36cb9209131e3225eaad6d7074b61da8a50abfa11c7a6a83cfd7773f3618af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:05:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.873462 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:05:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.891888 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fcdxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd9d66fb-a400-4810-aa7a-c81c9c24bd11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1341e3406e28b4e4309aaa18fba25c82ff489cc9cb44f232c4eb0f9a3762205c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:04:47Z\\\",\\\"message\\\":\\\"2025-12-03T16:04:01+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a1fbe999-84ce-42e2-80f8-36647b645314\\\\n2025-12-03T16:04:01+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a1fbe999-84ce-42e2-80f8-36647b645314 to /host/opt/cni/bin/\\\\n2025-12-03T16:04:01Z [verbose] multus-daemon started\\\\n2025-12-03T16:04:01Z [verbose] Readiness Indicator file check\\\\n2025-12-03T16:04:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6qw5j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fcdxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:05:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.905148 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:05:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.916643 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:05:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.921160 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.921188 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.921197 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.921212 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.921222 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:01Z","lastTransitionTime":"2025-12-03T16:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.929244 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27aeb142-d4e3-4827-ac1f-c34de6822b14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://509355988e03ff6660382aa4572bab44e3a122b3c67f9bcbabef6068939b2967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6htvc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9fjbr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:05:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.952127 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:00Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T16:04:36Z\\\",\\\"message\\\":\\\"dler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 16:04:36.686345 6741 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 16:04:36.686365 6741 factory.go:656] Stopping watch factory\\\\nI1203 16:04:36.686363 6741 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 16:04:36.686373 6741 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 16:04:36.686373 6741 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:36.686403 6741 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 16:04:36.686513 6741 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:36.686730 6741 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 16:04:36.686815 6741 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 16:04:36.686826 6741 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 16:04:36.686863 6741 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-m2lh2_openshift-ovn-kubernetes(4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-54d7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:04:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-m2lh2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:05:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.980521 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca0b3eec-84ee-423d-ae86-546de4ddfd9d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://427c4ea5ce8b52bdab1bb6f267d2a12056199db83590d5998fc161748161264c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4307ef354d034ea639e12cf4b3a9d80588cdefe30805aa434a2f3c920cabbb28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9cde717e32a20f4e9935b41cdc3a5d35a95646954785280609d5ab343851e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f7cd1d9f45e09425c7ae62de5d067dd211499fd508d7efc1c0372ce8e6b621b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f8b61c96a12c885b835790e20f50b94b257f392c9bbc104717f4bf6bb14137\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf7b0489cce8696026e6c955abe3188ce5acce768874caacf8f4655ad295034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72de226e60c9dabccdbda3e9bf98fe1e28448033fe2e36e27d779d024808f79d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2029e45c327bf297299a4b29f162c1aa242e11db775dd2a1e9070a9ef3e24610\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:05:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:05:01 crc kubenswrapper[4998]: I1203 16:05:01.996668 4998 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6dace3c-e5d0-43ab-9022-1c597f24e231\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:04:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T16:03:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T16:03:59Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1203 16:03:59.773604 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1203 16:03:59.773614 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1203 16:03:59.773618 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1203 16:03:59.773630 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1203 16:03:59.773632 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764777823\\\\\\\\\\\\\\\" (2025-12-03 16:03:43 +0000 UTC to 2026-01-02 16:03:44 +0000 UTC (now=2025-12-03 16:03:59.773600805 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773659 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-939894581/tls.crt::/tmp/serving-cert-939894581/tls.key\\\\\\\"\\\\nI1203 16:03:59.773742 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764777834\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764777834\\\\\\\\\\\\\\\" (2025-12-03 15:03:54 +0000 UTC to 2026-12-03 15:03:54 +0000 UTC (now=2025-12-03 16:03:59.773718168 +0000 UTC))\\\\\\\"\\\\nI1203 16:03:59.773760 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 16:03:59.773799 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1203 16:03:59.773817 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1203 16:03:59.774410 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T16:03:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T16:03:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T16:03:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T16:03:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:05:01Z is after 2025-08-24T17:21:41Z" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.023927 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.024003 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.024028 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.024056 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.024078 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:02Z","lastTransitionTime":"2025-12-03T16:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.052946 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.053005 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.053027 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.053051 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.053072 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:02Z","lastTransitionTime":"2025-12-03T16:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:02 crc kubenswrapper[4998]: E1203 16:05:02.075022 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:05:02Z is after 2025-08-24T17:21:41Z" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.080565 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.080645 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.080665 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.080690 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.080710 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:02Z","lastTransitionTime":"2025-12-03T16:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:02 crc kubenswrapper[4998]: E1203 16:05:02.101347 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:05:02Z is after 2025-08-24T17:21:41Z" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.106123 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.106181 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.106203 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.106222 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.106238 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:02Z","lastTransitionTime":"2025-12-03T16:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:02 crc kubenswrapper[4998]: E1203 16:05:02.124931 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:05:02Z is after 2025-08-24T17:21:41Z" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.129310 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.129346 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.129358 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.129374 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.129385 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:02Z","lastTransitionTime":"2025-12-03T16:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:02 crc kubenswrapper[4998]: E1203 16:05:02.144055 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:05:02Z is after 2025-08-24T17:21:41Z" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.147538 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.147580 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.147594 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.147610 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.147621 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:02Z","lastTransitionTime":"2025-12-03T16:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:02 crc kubenswrapper[4998]: E1203 16:05:02.160191 4998 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T16:05:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ef69ca4e-e477-4474-a354-15f3061d008d\\\",\\\"systemUUID\\\":\\\"0d8400b0-b239-4c8c-8060-4c5edd34a62b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T16:05:02Z is after 2025-08-24T17:21:41Z" Dec 03 16:05:02 crc kubenswrapper[4998]: E1203 16:05:02.160591 4998 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.162967 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.163027 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.163043 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.163057 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.163075 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:02Z","lastTransitionTime":"2025-12-03T16:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.266257 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.266535 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.266668 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.266850 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.266997 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:02Z","lastTransitionTime":"2025-12-03T16:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.369690 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.369780 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.369800 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.369825 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.369844 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:02Z","lastTransitionTime":"2025-12-03T16:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.473013 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.473063 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.473072 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.473088 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.473096 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:02Z","lastTransitionTime":"2025-12-03T16:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.575430 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.575491 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.575514 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.575547 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.575570 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:02Z","lastTransitionTime":"2025-12-03T16:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.676885 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:05:02 crc kubenswrapper[4998]: E1203 16:05:02.677135 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.676903 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:05:02 crc kubenswrapper[4998]: E1203 16:05:02.677273 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.676885 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:05:02 crc kubenswrapper[4998]: E1203 16:05:02.677366 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.678629 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.678789 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.678887 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.679030 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.679167 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:02Z","lastTransitionTime":"2025-12-03T16:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.782180 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.782228 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.782248 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.782271 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.782287 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:02Z","lastTransitionTime":"2025-12-03T16:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.884377 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.884435 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.884451 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.884480 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.884493 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:02Z","lastTransitionTime":"2025-12-03T16:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.987105 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.987167 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.987179 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.987199 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:02 crc kubenswrapper[4998]: I1203 16:05:02.987215 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:02Z","lastTransitionTime":"2025-12-03T16:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.090065 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.090122 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.090133 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.090150 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.090163 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:03Z","lastTransitionTime":"2025-12-03T16:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.193160 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.193212 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.193224 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.193245 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.193261 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:03Z","lastTransitionTime":"2025-12-03T16:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.296338 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.296407 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.296430 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.296459 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.296480 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:03Z","lastTransitionTime":"2025-12-03T16:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.398943 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.399011 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.399023 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.399044 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.399056 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:03Z","lastTransitionTime":"2025-12-03T16:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.501395 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.501441 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.501452 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.501470 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.501482 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:03Z","lastTransitionTime":"2025-12-03T16:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.604889 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.604937 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.604948 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.604968 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.604982 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:03Z","lastTransitionTime":"2025-12-03T16:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.677482 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:05:03 crc kubenswrapper[4998]: E1203 16:05:03.678422 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.678490 4998 scope.go:117] "RemoveContainer" containerID="05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.708275 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.708329 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.708346 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.708372 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.708392 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:03Z","lastTransitionTime":"2025-12-03T16:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.812553 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.812627 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.812647 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.812680 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.812702 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:03Z","lastTransitionTime":"2025-12-03T16:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.916546 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.916624 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.916659 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.916690 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:03 crc kubenswrapper[4998]: I1203 16:05:03.916711 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:03Z","lastTransitionTime":"2025-12-03T16:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.020456 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.020540 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.020574 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.020608 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.020631 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:04Z","lastTransitionTime":"2025-12-03T16:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.123518 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.123593 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.123627 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.123656 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.123677 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:04Z","lastTransitionTime":"2025-12-03T16:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.227313 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.227360 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.227371 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.227388 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.227401 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:04Z","lastTransitionTime":"2025-12-03T16:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.330605 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.330667 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.330689 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.330720 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.330951 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:04Z","lastTransitionTime":"2025-12-03T16:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.434982 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.435456 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.435652 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.435898 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.436129 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:04Z","lastTransitionTime":"2025-12-03T16:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.540113 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.540182 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.540202 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.540230 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.540250 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:04Z","lastTransitionTime":"2025-12-03T16:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.593162 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:04 crc kubenswrapper[4998]: E1203 16:05:04.593350 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:06:08.593313876 +0000 UTC m=+147.205014129 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.593944 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:05:04 crc kubenswrapper[4998]: E1203 16:05:04.594203 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:05:04 crc kubenswrapper[4998]: E1203 16:05:04.594264 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.594210 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.594371 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.594424 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:05:04 crc kubenswrapper[4998]: E1203 16:05:04.594292 4998 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:05:04 crc kubenswrapper[4998]: E1203 16:05:04.594518 4998 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:05:04 crc kubenswrapper[4998]: E1203 16:05:04.594570 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 16:06:08.594538473 +0000 UTC m=+147.206238726 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:05:04 crc kubenswrapper[4998]: E1203 16:05:04.594600 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:06:08.594587224 +0000 UTC m=+147.206287487 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 16:05:04 crc kubenswrapper[4998]: E1203 16:05:04.594603 4998 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:05:04 crc kubenswrapper[4998]: E1203 16:05:04.594681 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 16:06:08.594658015 +0000 UTC m=+147.206358278 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 16:05:04 crc kubenswrapper[4998]: E1203 16:05:04.594949 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 16:05:04 crc kubenswrapper[4998]: E1203 16:05:04.595025 4998 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 16:05:04 crc kubenswrapper[4998]: E1203 16:05:04.595049 4998 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:05:04 crc kubenswrapper[4998]: E1203 16:05:04.595185 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 16:06:08.595159197 +0000 UTC m=+147.206859460 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.643364 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.643463 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.643489 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.643526 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.643549 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:04Z","lastTransitionTime":"2025-12-03T16:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.677420 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.677474 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.677419 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:05:04 crc kubenswrapper[4998]: E1203 16:05:04.677921 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:05:04 crc kubenswrapper[4998]: E1203 16:05:04.678022 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:05:04 crc kubenswrapper[4998]: E1203 16:05:04.678124 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.748267 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.748344 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.748368 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.748399 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.748422 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:04Z","lastTransitionTime":"2025-12-03T16:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.851437 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.851480 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.851491 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.851509 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.851523 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:04Z","lastTransitionTime":"2025-12-03T16:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.954609 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.954948 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.954964 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.954982 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:04 crc kubenswrapper[4998]: I1203 16:05:04.954993 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:04Z","lastTransitionTime":"2025-12-03T16:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.058188 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.058232 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.058244 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.058262 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.058275 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:05Z","lastTransitionTime":"2025-12-03T16:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.161102 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.161185 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.161214 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.161245 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.161267 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:05Z","lastTransitionTime":"2025-12-03T16:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.185226 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m2lh2_4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9/ovnkube-controller/2.log" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.189409 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" event={"ID":"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9","Type":"ContainerStarted","Data":"5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df"} Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.190218 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.264256 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.264332 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.264352 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.264378 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.264395 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:05Z","lastTransitionTime":"2025-12-03T16:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.274071 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-9srmx" podStartSLOduration=66.274042535 podStartE2EDuration="1m6.274042535s" podCreationTimestamp="2025-12-03 16:03:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:05.256507756 +0000 UTC m=+83.868208009" watchObservedRunningTime="2025-12-03 16:05:05.274042535 +0000 UTC m=+83.885742788" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.274283 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-74tc7" podStartSLOduration=66.274267059 podStartE2EDuration="1m6.274267059s" podCreationTimestamp="2025-12-03 16:03:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:05.273867161 +0000 UTC m=+83.885567444" watchObservedRunningTime="2025-12-03 16:05:05.274267059 +0000 UTC m=+83.885967323" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.313307 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=33.313283924 podStartE2EDuration="33.313283924s" podCreationTimestamp="2025-12-03 16:04:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:05.313250063 +0000 UTC m=+83.924950316" watchObservedRunningTime="2025-12-03 16:05:05.313283924 +0000 UTC m=+83.924984157" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.350522 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=15.350492928 podStartE2EDuration="15.350492928s" podCreationTimestamp="2025-12-03 16:04:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:05.32712182 +0000 UTC m=+83.938822063" watchObservedRunningTime="2025-12-03 16:05:05.350492928 +0000 UTC m=+83.962193191" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.351018 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-qt6sw" podStartSLOduration=66.351007759 podStartE2EDuration="1m6.351007759s" podCreationTimestamp="2025-12-03 16:03:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:05.350154191 +0000 UTC m=+83.961854434" watchObservedRunningTime="2025-12-03 16:05:05.351007759 +0000 UTC m=+83.962708032" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.366286 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.366326 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.366337 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.366352 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.366364 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:05Z","lastTransitionTime":"2025-12-03T16:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.391893 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47k9d" podStartSLOduration=64.391868555 podStartE2EDuration="1m4.391868555s" podCreationTimestamp="2025-12-03 16:04:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:05.365391088 +0000 UTC m=+83.977091331" watchObservedRunningTime="2025-12-03 16:05:05.391868555 +0000 UTC m=+84.003568788" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.392241 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=59.392233853 podStartE2EDuration="59.392233853s" podCreationTimestamp="2025-12-03 16:04:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:05.39167809 +0000 UTC m=+84.003378313" watchObservedRunningTime="2025-12-03 16:05:05.392233853 +0000 UTC m=+84.003934086" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.451144 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-fcdxh" podStartSLOduration=66.451127877 podStartE2EDuration="1m6.451127877s" podCreationTimestamp="2025-12-03 16:03:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:05.450747129 +0000 UTC m=+84.062447362" watchObservedRunningTime="2025-12-03 16:05:05.451127877 +0000 UTC m=+84.062828100" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.468742 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.468816 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.468832 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.468854 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.468871 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:05Z","lastTransitionTime":"2025-12-03T16:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.502857 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podStartSLOduration=66.502841283 podStartE2EDuration="1m6.502841283s" podCreationTimestamp="2025-12-03 16:03:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:05.501863691 +0000 UTC m=+84.113563924" watchObservedRunningTime="2025-12-03 16:05:05.502841283 +0000 UTC m=+84.114541506" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.538796 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" podStartSLOduration=65.538777369 podStartE2EDuration="1m5.538777369s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:05.538189506 +0000 UTC m=+84.149889749" watchObservedRunningTime="2025-12-03 16:05:05.538777369 +0000 UTC m=+84.150477592" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.564212 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=62.564195042 podStartE2EDuration="1m2.564195042s" podCreationTimestamp="2025-12-03 16:04:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:05.563062257 +0000 UTC m=+84.174762490" watchObservedRunningTime="2025-12-03 16:05:05.564195042 +0000 UTC m=+84.175895265" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.571415 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.571457 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.571466 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.571483 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.571494 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:05Z","lastTransitionTime":"2025-12-03T16:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.674144 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.674183 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.674192 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.674207 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.674219 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:05Z","lastTransitionTime":"2025-12-03T16:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.677421 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:05:05 crc kubenswrapper[4998]: E1203 16:05:05.677525 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.776351 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.776382 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.776390 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.776403 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.776411 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:05Z","lastTransitionTime":"2025-12-03T16:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.878195 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.878251 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.878268 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.878290 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.878306 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:05Z","lastTransitionTime":"2025-12-03T16:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.932413 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=65.932382927 podStartE2EDuration="1m5.932382927s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:05.581615428 +0000 UTC m=+84.193315661" watchObservedRunningTime="2025-12-03 16:05:05.932382927 +0000 UTC m=+84.544083190" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.934042 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-k8ptd"] Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.981219 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.981275 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.981284 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.981300 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:05 crc kubenswrapper[4998]: I1203 16:05:05.981316 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:05Z","lastTransitionTime":"2025-12-03T16:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.084083 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.084115 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.084127 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.084143 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.084154 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:06Z","lastTransitionTime":"2025-12-03T16:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.186370 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.186415 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.186427 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.186442 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.186453 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:06Z","lastTransitionTime":"2025-12-03T16:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.192047 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:05:06 crc kubenswrapper[4998]: E1203 16:05:06.192269 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.289843 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.289904 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.289924 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.289954 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.289973 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:06Z","lastTransitionTime":"2025-12-03T16:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.392950 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.393028 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.393053 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.393085 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.393106 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:06Z","lastTransitionTime":"2025-12-03T16:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.496241 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.496309 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.496331 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.496354 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.496387 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:06Z","lastTransitionTime":"2025-12-03T16:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.600329 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.600394 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.600411 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.600437 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.600454 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:06Z","lastTransitionTime":"2025-12-03T16:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.677211 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.677336 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.677902 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:05:06 crc kubenswrapper[4998]: E1203 16:05:06.678111 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 16:05:06 crc kubenswrapper[4998]: E1203 16:05:06.678200 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 16:05:06 crc kubenswrapper[4998]: E1203 16:05:06.678303 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.703416 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.703480 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.703497 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.703522 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.703540 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:06Z","lastTransitionTime":"2025-12-03T16:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.807082 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.807149 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.807166 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.807189 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.807205 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:06Z","lastTransitionTime":"2025-12-03T16:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.910349 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.910415 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.910439 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.910469 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:06 crc kubenswrapper[4998]: I1203 16:05:06.910493 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:06Z","lastTransitionTime":"2025-12-03T16:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.014336 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.014413 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.014431 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.014457 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.014478 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:07Z","lastTransitionTime":"2025-12-03T16:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.117112 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.117161 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.117177 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.117193 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.117203 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:07Z","lastTransitionTime":"2025-12-03T16:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.218923 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.218960 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.218970 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.218985 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.218994 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:07Z","lastTransitionTime":"2025-12-03T16:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.322593 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.322660 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.322681 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.322705 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.322722 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:07Z","lastTransitionTime":"2025-12-03T16:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.425180 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.425311 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.425336 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.425364 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.425383 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:07Z","lastTransitionTime":"2025-12-03T16:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.527663 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.527732 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.527748 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.527798 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.527817 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:07Z","lastTransitionTime":"2025-12-03T16:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.631294 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.631341 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.631357 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.631381 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.631397 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:07Z","lastTransitionTime":"2025-12-03T16:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.677340 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:05:07 crc kubenswrapper[4998]: E1203 16:05:07.677734 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k8ptd" podUID="2914eb52-522e-4789-a988-b15875755144" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.734150 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.734214 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.734231 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.734255 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.734278 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:07Z","lastTransitionTime":"2025-12-03T16:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.837398 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.837470 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.837494 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.837519 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.837536 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:07Z","lastTransitionTime":"2025-12-03T16:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.940378 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.940436 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.940452 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.940475 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 16:05:07 crc kubenswrapper[4998]: I1203 16:05:07.940491 4998 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T16:05:07Z","lastTransitionTime":"2025-12-03T16:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.044019 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.044066 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.044082 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.044102 4998 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.044205 4998 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.089469 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2kprn"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.090034 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hwftp"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.092919 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-6jzdp"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.092981 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.094067 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jzdp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.095164 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.096938 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9cdmt"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.097228 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.097728 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9cdmt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.109367 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.115737 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-4rq7l"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.116488 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.117319 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-4rq7l" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.124926 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.125160 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.125855 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.133264 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a556e01d-7f37-46db-ac77-6a681cad9e7e-config\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.133314 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20e7b7c5-1480-498e-ab4d-2e335da52dd3-config\") pod \"machine-approver-56656f9798-6jzdp\" (UID: \"20e7b7c5-1480-498e-ab4d-2e335da52dd3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jzdp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.133375 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a556e01d-7f37-46db-ac77-6a681cad9e7e-etcd-serving-ca\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.133415 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-542jg\" (UniqueName: \"kubernetes.io/projected/b46224fb-c151-4ba9-87c7-85e0ed6e11bc-kube-api-access-542jg\") pod \"apiserver-7bbb656c7d-bbss6\" (UID: \"b46224fb-c151-4ba9-87c7-85e0ed6e11bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.133457 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/94922a63-40d9-4944-9d95-f860967ea5b3-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-9cdmt\" (UID: \"94922a63-40d9-4944-9d95-f860967ea5b3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9cdmt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.133511 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b46224fb-c151-4ba9-87c7-85e0ed6e11bc-audit-dir\") pod \"apiserver-7bbb656c7d-bbss6\" (UID: \"b46224fb-c151-4ba9-87c7-85e0ed6e11bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.133559 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c2ff7dad-a4da-4243-ab1e-21b47230504d-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2kprn\" (UID: \"c2ff7dad-a4da-4243-ab1e-21b47230504d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.133584 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a556e01d-7f37-46db-ac77-6a681cad9e7e-node-pullsecrets\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.133618 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxdjj\" (UniqueName: \"kubernetes.io/projected/c2ff7dad-a4da-4243-ab1e-21b47230504d-kube-api-access-bxdjj\") pod \"controller-manager-879f6c89f-2kprn\" (UID: \"c2ff7dad-a4da-4243-ab1e-21b47230504d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.133643 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/a556e01d-7f37-46db-ac77-6a681cad9e7e-audit\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.133678 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txxpv\" (UniqueName: \"kubernetes.io/projected/20e7b7c5-1480-498e-ab4d-2e335da52dd3-kube-api-access-txxpv\") pod \"machine-approver-56656f9798-6jzdp\" (UID: \"20e7b7c5-1480-498e-ab4d-2e335da52dd3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jzdp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.134350 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/a556e01d-7f37-46db-ac77-6a681cad9e7e-image-import-ca\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.134403 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b46224fb-c151-4ba9-87c7-85e0ed6e11bc-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-bbss6\" (UID: \"b46224fb-c151-4ba9-87c7-85e0ed6e11bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.134438 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b46224fb-c151-4ba9-87c7-85e0ed6e11bc-encryption-config\") pod \"apiserver-7bbb656c7d-bbss6\" (UID: \"b46224fb-c151-4ba9-87c7-85e0ed6e11bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.134478 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4dj2\" (UniqueName: \"kubernetes.io/projected/94922a63-40d9-4944-9d95-f860967ea5b3-kube-api-access-d4dj2\") pod \"cluster-samples-operator-665b6dd947-9cdmt\" (UID: \"94922a63-40d9-4944-9d95-f860967ea5b3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9cdmt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.134505 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a556e01d-7f37-46db-ac77-6a681cad9e7e-serving-cert\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.134531 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a556e01d-7f37-46db-ac77-6a681cad9e7e-encryption-config\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.134564 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b46224fb-c151-4ba9-87c7-85e0ed6e11bc-etcd-client\") pod \"apiserver-7bbb656c7d-bbss6\" (UID: \"b46224fb-c151-4ba9-87c7-85e0ed6e11bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.134599 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2ff7dad-a4da-4243-ab1e-21b47230504d-config\") pod \"controller-manager-879f6c89f-2kprn\" (UID: \"c2ff7dad-a4da-4243-ab1e-21b47230504d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.134624 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a556e01d-7f37-46db-ac77-6a681cad9e7e-etcd-client\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.134673 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c2ff7dad-a4da-4243-ab1e-21b47230504d-client-ca\") pod \"controller-manager-879f6c89f-2kprn\" (UID: \"c2ff7dad-a4da-4243-ab1e-21b47230504d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.134710 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/20e7b7c5-1480-498e-ab4d-2e335da52dd3-machine-approver-tls\") pod \"machine-approver-56656f9798-6jzdp\" (UID: \"20e7b7c5-1480-498e-ab4d-2e335da52dd3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jzdp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.134745 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b46224fb-c151-4ba9-87c7-85e0ed6e11bc-serving-cert\") pod \"apiserver-7bbb656c7d-bbss6\" (UID: \"b46224fb-c151-4ba9-87c7-85e0ed6e11bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.134806 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a556e01d-7f37-46db-ac77-6a681cad9e7e-audit-dir\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.134836 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/20e7b7c5-1480-498e-ab4d-2e335da52dd3-auth-proxy-config\") pod \"machine-approver-56656f9798-6jzdp\" (UID: \"20e7b7c5-1480-498e-ab4d-2e335da52dd3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jzdp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.134873 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b46224fb-c151-4ba9-87c7-85e0ed6e11bc-audit-policies\") pod \"apiserver-7bbb656c7d-bbss6\" (UID: \"b46224fb-c151-4ba9-87c7-85e0ed6e11bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.134922 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a556e01d-7f37-46db-ac77-6a681cad9e7e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.134955 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b46224fb-c151-4ba9-87c7-85e0ed6e11bc-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-bbss6\" (UID: \"b46224fb-c151-4ba9-87c7-85e0ed6e11bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.134981 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2ff7dad-a4da-4243-ab1e-21b47230504d-serving-cert\") pod \"controller-manager-879f6c89f-2kprn\" (UID: \"c2ff7dad-a4da-4243-ab1e-21b47230504d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.135005 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbnql\" (UniqueName: \"kubernetes.io/projected/a556e01d-7f37-46db-ac77-6a681cad9e7e-kube-api-access-sbnql\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.135610 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.135714 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.136556 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.136601 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.136917 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.137052 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.137387 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.137485 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.137517 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.137853 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.156678 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.157562 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.157672 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.157897 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.158046 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.158195 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.158306 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.159307 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.159464 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.159842 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.159910 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.159974 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.160029 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.160077 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.160137 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.160196 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.160227 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.160321 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.160335 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.160408 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.160435 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.160527 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.160555 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.160633 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.160665 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.160728 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.160783 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.160831 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-8ft9r"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.161467 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8ft9r" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.162615 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.162732 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.163480 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.163905 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nsq9q"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.164363 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nsq9q" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.166376 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.166687 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.166791 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wk6qg"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.167362 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.168344 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.171052 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.171468 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.172644 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.172995 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.175998 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-ggw6k"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.173265 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.173296 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.176556 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.176828 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.176837 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.176905 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.176110 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wk6qg" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.182373 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.185300 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qvmlb"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.185518 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-kg4lw"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.185851 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctpbx"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.185856 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.186083 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-kk8mb"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.186232 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.186320 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-54vrz"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.186368 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.186667 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.186855 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-qvmlb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.186930 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kg4lw" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.187277 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-ggw6k" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.187386 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctpbx" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.188494 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.200023 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-ppw5v"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.201019 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-5c2lp"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.201139 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.201604 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.201742 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-5c2lp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.206410 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5wc9c"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.209478 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-ppw5v" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.201748 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.209628 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.202043 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.202089 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.202151 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.202196 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.202342 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.202400 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.202519 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.202571 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.212376 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-54vrz" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.202868 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.204939 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.205145 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.205239 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.205927 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.206437 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.206576 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.209235 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.215775 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.216275 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.223260 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.223395 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.223667 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.223864 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8lmkd"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.223927 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.224176 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.224372 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-b5hpf"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.224563 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5wc9c" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.224932 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-r2gbh"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.225286 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8lmkd" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.225449 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-r2gbh" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.225491 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rrj4k"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.225655 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.226153 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rrj4k" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.226855 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-55d4c"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.227342 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.227898 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xd7m2"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.228200 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xd7m2" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.231659 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.231989 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-hnvvp"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.232659 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-dc55h"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.233280 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dc55h" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.233664 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-hnvvp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.234302 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.236412 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.236657 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.236782 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237029 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/a556e01d-7f37-46db-ac77-6a681cad9e7e-audit\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237065 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a1b6dae-ff3d-4bfb-a811-98866c9c2b59-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-54vrz\" (UID: \"0a1b6dae-ff3d-4bfb-a811-98866c9c2b59\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-54vrz" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237093 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxdjj\" (UniqueName: \"kubernetes.io/projected/c2ff7dad-a4da-4243-ab1e-21b47230504d-kube-api-access-bxdjj\") pod \"controller-manager-879f6c89f-2kprn\" (UID: \"c2ff7dad-a4da-4243-ab1e-21b47230504d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237117 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2r7z\" (UniqueName: \"kubernetes.io/projected/56ad3380-4861-49d1-8758-7b1e27f74560-kube-api-access-n2r7z\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237150 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09d80f4a-297b-4000-8fcd-c586b5267339-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-nsq9q\" (UID: \"09d80f4a-297b-4000-8fcd-c586b5267339\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nsq9q" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237174 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/03021110-7732-4dff-a5e3-f481e8b7c0d6-console-serving-cert\") pod \"console-f9d7485db-kk8mb\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237197 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/235934fb-ec04-4784-b30c-5c80dfa325ad-service-ca-bundle\") pod \"authentication-operator-69f744f599-5c2lp\" (UID: \"235934fb-ec04-4784-b30c-5c80dfa325ad\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5c2lp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237226 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txxpv\" (UniqueName: \"kubernetes.io/projected/20e7b7c5-1480-498e-ab4d-2e335da52dd3-kube-api-access-txxpv\") pod \"machine-approver-56656f9798-6jzdp\" (UID: \"20e7b7c5-1480-498e-ab4d-2e335da52dd3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jzdp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237250 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1ff9891-7017-41a2-b7c0-183caa783671-serving-cert\") pod \"console-operator-58897d9998-qvmlb\" (UID: \"e1ff9891-7017-41a2-b7c0-183caa783671\") " pod="openshift-console-operator/console-operator-58897d9998-qvmlb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237274 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/892e64fc-db6b-4410-9b4c-3ff78cc82074-images\") pod \"machine-api-operator-5694c8668f-ggw6k\" (UID: \"892e64fc-db6b-4410-9b4c-3ff78cc82074\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ggw6k" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237296 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/56ad3380-4861-49d1-8758-7b1e27f74560-audit-policies\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237320 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/a556e01d-7f37-46db-ac77-6a681cad9e7e-image-import-ca\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237342 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e569f94d-9231-4277-af77-f97c23fcbabe-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-wk6qg\" (UID: \"e569f94d-9231-4277-af77-f97c23fcbabe\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wk6qg" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237364 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e1ff9891-7017-41a2-b7c0-183caa783671-trusted-ca\") pod \"console-operator-58897d9998-qvmlb\" (UID: \"e1ff9891-7017-41a2-b7c0-183caa783671\") " pod="openshift-console-operator/console-operator-58897d9998-qvmlb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237387 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b46224fb-c151-4ba9-87c7-85e0ed6e11bc-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-bbss6\" (UID: \"b46224fb-c151-4ba9-87c7-85e0ed6e11bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237408 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b46224fb-c151-4ba9-87c7-85e0ed6e11bc-encryption-config\") pod \"apiserver-7bbb656c7d-bbss6\" (UID: \"b46224fb-c151-4ba9-87c7-85e0ed6e11bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237431 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/84e61bdd-8de2-4cd1-93a2-b43d802fbea3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-ctpbx\" (UID: \"84e61bdd-8de2-4cd1-93a2-b43d802fbea3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctpbx" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237470 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4dj2\" (UniqueName: \"kubernetes.io/projected/94922a63-40d9-4944-9d95-f860967ea5b3-kube-api-access-d4dj2\") pod \"cluster-samples-operator-665b6dd947-9cdmt\" (UID: \"94922a63-40d9-4944-9d95-f860967ea5b3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9cdmt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237494 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/22c37be4-5e07-4638-aa01-04ee06c463d7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-r2gbh\" (UID: \"22c37be4-5e07-4638-aa01-04ee06c463d7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-r2gbh" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237518 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a556e01d-7f37-46db-ac77-6a681cad9e7e-serving-cert\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237539 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09d80f4a-297b-4000-8fcd-c586b5267339-config\") pod \"openshift-apiserver-operator-796bbdcf4f-nsq9q\" (UID: \"09d80f4a-297b-4000-8fcd-c586b5267339\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nsq9q" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237563 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237585 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e569f94d-9231-4277-af77-f97c23fcbabe-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-wk6qg\" (UID: \"e569f94d-9231-4277-af77-f97c23fcbabe\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wk6qg" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237606 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a1b6dae-ff3d-4bfb-a811-98866c9c2b59-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-54vrz\" (UID: \"0a1b6dae-ff3d-4bfb-a811-98866c9c2b59\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-54vrz" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237629 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68597ef9-34cf-45c4-86c7-153f9f4d9363-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5wc9c\" (UID: \"68597ef9-34cf-45c4-86c7-153f9f4d9363\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5wc9c" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237656 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84e61bdd-8de2-4cd1-93a2-b43d802fbea3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-ctpbx\" (UID: \"84e61bdd-8de2-4cd1-93a2-b43d802fbea3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctpbx" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237678 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68597ef9-34cf-45c4-86c7-153f9f4d9363-config\") pod \"kube-controller-manager-operator-78b949d7b-5wc9c\" (UID: \"68597ef9-34cf-45c4-86c7-153f9f4d9363\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5wc9c" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237697 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c71a13b3-a727-4b88-9be8-954a62dc7c86-bound-sa-token\") pod \"ingress-operator-5b745b69d9-8ft9r\" (UID: \"c71a13b3-a727-4b88-9be8-954a62dc7c86\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8ft9r" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237723 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a556e01d-7f37-46db-ac77-6a681cad9e7e-encryption-config\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237746 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b46224fb-c151-4ba9-87c7-85e0ed6e11bc-etcd-client\") pod \"apiserver-7bbb656c7d-bbss6\" (UID: \"b46224fb-c151-4ba9-87c7-85e0ed6e11bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237789 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/57c54ccc-a846-4c00-8318-5d5331c70e70-metrics-tls\") pod \"dns-operator-744455d44c-4rq7l\" (UID: \"57c54ccc-a846-4c00-8318-5d5331c70e70\") " pod="openshift-dns-operator/dns-operator-744455d44c-4rq7l" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237814 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8hgn\" (UniqueName: \"kubernetes.io/projected/57c54ccc-a846-4c00-8318-5d5331c70e70-kube-api-access-q8hgn\") pod \"dns-operator-744455d44c-4rq7l\" (UID: \"57c54ccc-a846-4c00-8318-5d5331c70e70\") " pod="openshift-dns-operator/dns-operator-744455d44c-4rq7l" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237841 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237880 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzn2p\" (UniqueName: \"kubernetes.io/projected/6024dcd8-96d6-4c42-9670-42e3608cd7ee-kube-api-access-tzn2p\") pod \"route-controller-manager-6576b87f9c-tvqq9\" (UID: \"6024dcd8-96d6-4c42-9670-42e3608cd7ee\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237911 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7a975a4d-12df-4c73-b644-ff281b3c3390-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8lmkd\" (UID: \"7a975a4d-12df-4c73-b644-ff281b3c3390\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8lmkd" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237937 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237965 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2ff7dad-a4da-4243-ab1e-21b47230504d-config\") pod \"controller-manager-879f6c89f-2kprn\" (UID: \"c2ff7dad-a4da-4243-ab1e-21b47230504d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.237988 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c71a13b3-a727-4b88-9be8-954a62dc7c86-trusted-ca\") pod \"ingress-operator-5b745b69d9-8ft9r\" (UID: \"c71a13b3-a727-4b88-9be8-954a62dc7c86\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8ft9r" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238010 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238037 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqpn4\" (UniqueName: \"kubernetes.io/projected/09d80f4a-297b-4000-8fcd-c586b5267339-kube-api-access-gqpn4\") pod \"openshift-apiserver-operator-796bbdcf4f-nsq9q\" (UID: \"09d80f4a-297b-4000-8fcd-c586b5267339\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nsq9q" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238099 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7r6j\" (UniqueName: \"kubernetes.io/projected/e1ff9891-7017-41a2-b7c0-183caa783671-kube-api-access-g7r6j\") pod \"console-operator-58897d9998-qvmlb\" (UID: \"e1ff9891-7017-41a2-b7c0-183caa783671\") " pod="openshift-console-operator/console-operator-58897d9998-qvmlb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238127 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238149 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/7a975a4d-12df-4c73-b644-ff281b3c3390-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8lmkd\" (UID: \"7a975a4d-12df-4c73-b644-ff281b3c3390\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8lmkd" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238185 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238221 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238251 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a556e01d-7f37-46db-ac77-6a681cad9e7e-etcd-client\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238282 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68597ef9-34cf-45c4-86c7-153f9f4d9363-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5wc9c\" (UID: \"68597ef9-34cf-45c4-86c7-153f9f4d9363\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5wc9c" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238308 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjd2p\" (UniqueName: \"kubernetes.io/projected/03021110-7732-4dff-a5e3-f481e8b7c0d6-kube-api-access-pjd2p\") pod \"console-f9d7485db-kk8mb\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238343 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zk89t\" (UniqueName: \"kubernetes.io/projected/892e64fc-db6b-4410-9b4c-3ff78cc82074-kube-api-access-zk89t\") pod \"machine-api-operator-5694c8668f-ggw6k\" (UID: \"892e64fc-db6b-4410-9b4c-3ff78cc82074\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ggw6k" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238366 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkxbv\" (UniqueName: \"kubernetes.io/projected/0a1b6dae-ff3d-4bfb-a811-98866c9c2b59-kube-api-access-hkxbv\") pod \"openshift-controller-manager-operator-756b6f6bc6-54vrz\" (UID: \"0a1b6dae-ff3d-4bfb-a811-98866c9c2b59\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-54vrz" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238387 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/03021110-7732-4dff-a5e3-f481e8b7c0d6-oauth-serving-cert\") pod \"console-f9d7485db-kk8mb\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238410 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c2ff7dad-a4da-4243-ab1e-21b47230504d-client-ca\") pod \"controller-manager-879f6c89f-2kprn\" (UID: \"c2ff7dad-a4da-4243-ab1e-21b47230504d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238431 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/03021110-7732-4dff-a5e3-f481e8b7c0d6-console-config\") pod \"console-f9d7485db-kk8mb\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238460 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1ff9891-7017-41a2-b7c0-183caa783671-config\") pod \"console-operator-58897d9998-qvmlb\" (UID: \"e1ff9891-7017-41a2-b7c0-183caa783671\") " pod="openshift-console-operator/console-operator-58897d9998-qvmlb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238483 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/ac8c6176-0450-4050-9094-003067cd40b1-available-featuregates\") pod \"openshift-config-operator-7777fb866f-kg4lw\" (UID: \"ac8c6176-0450-4050-9094-003067cd40b1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kg4lw" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238502 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/56ad3380-4861-49d1-8758-7b1e27f74560-audit-dir\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238529 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/20e7b7c5-1480-498e-ab4d-2e335da52dd3-machine-approver-tls\") pod \"machine-approver-56656f9798-6jzdp\" (UID: \"20e7b7c5-1480-498e-ab4d-2e335da52dd3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jzdp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238555 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b46224fb-c151-4ba9-87c7-85e0ed6e11bc-serving-cert\") pod \"apiserver-7bbb656c7d-bbss6\" (UID: \"b46224fb-c151-4ba9-87c7-85e0ed6e11bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238579 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/892e64fc-db6b-4410-9b4c-3ff78cc82074-config\") pod \"machine-api-operator-5694c8668f-ggw6k\" (UID: \"892e64fc-db6b-4410-9b4c-3ff78cc82074\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ggw6k" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238601 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c71a13b3-a727-4b88-9be8-954a62dc7c86-metrics-tls\") pod \"ingress-operator-5b745b69d9-8ft9r\" (UID: \"c71a13b3-a727-4b88-9be8-954a62dc7c86\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8ft9r" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238621 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/03021110-7732-4dff-a5e3-f481e8b7c0d6-console-oauth-config\") pod \"console-f9d7485db-kk8mb\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238645 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a556e01d-7f37-46db-ac77-6a681cad9e7e-audit-dir\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238664 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/20e7b7c5-1480-498e-ab4d-2e335da52dd3-auth-proxy-config\") pod \"machine-approver-56656f9798-6jzdp\" (UID: \"20e7b7c5-1480-498e-ab4d-2e335da52dd3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jzdp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238685 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b46224fb-c151-4ba9-87c7-85e0ed6e11bc-audit-policies\") pod \"apiserver-7bbb656c7d-bbss6\" (UID: \"b46224fb-c151-4ba9-87c7-85e0ed6e11bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238705 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7a975a4d-12df-4c73-b644-ff281b3c3390-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8lmkd\" (UID: \"7a975a4d-12df-4c73-b644-ff281b3c3390\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8lmkd" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.238725 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84e61bdd-8de2-4cd1-93a2-b43d802fbea3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-ctpbx\" (UID: \"84e61bdd-8de2-4cd1-93a2-b43d802fbea3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctpbx" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.239272 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.239555 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.239813 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e569f94d-9231-4277-af77-f97c23fcbabe-config\") pod \"kube-apiserver-operator-766d6c64bb-wk6qg\" (UID: \"e569f94d-9231-4277-af77-f97c23fcbabe\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wk6qg" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.239850 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxtbv\" (UniqueName: \"kubernetes.io/projected/235934fb-ec04-4784-b30c-5c80dfa325ad-kube-api-access-bxtbv\") pod \"authentication-operator-69f744f599-5c2lp\" (UID: \"235934fb-ec04-4784-b30c-5c80dfa325ad\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5c2lp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.239875 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.239904 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a556e01d-7f37-46db-ac77-6a681cad9e7e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.239929 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b46224fb-c151-4ba9-87c7-85e0ed6e11bc-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-bbss6\" (UID: \"b46224fb-c151-4ba9-87c7-85e0ed6e11bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.239954 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/892e64fc-db6b-4410-9b4c-3ff78cc82074-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-ggw6k\" (UID: \"892e64fc-db6b-4410-9b4c-3ff78cc82074\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ggw6k" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.239981 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/235934fb-ec04-4784-b30c-5c80dfa325ad-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-5c2lp\" (UID: \"235934fb-ec04-4784-b30c-5c80dfa325ad\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5c2lp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240006 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2ff7dad-a4da-4243-ab1e-21b47230504d-serving-cert\") pod \"controller-manager-879f6c89f-2kprn\" (UID: \"c2ff7dad-a4da-4243-ab1e-21b47230504d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240031 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbnql\" (UniqueName: \"kubernetes.io/projected/a556e01d-7f37-46db-ac77-6a681cad9e7e-kube-api-access-sbnql\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240056 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wst5d\" (UniqueName: \"kubernetes.io/projected/22c37be4-5e07-4638-aa01-04ee06c463d7-kube-api-access-wst5d\") pod \"multus-admission-controller-857f4d67dd-r2gbh\" (UID: \"22c37be4-5e07-4638-aa01-04ee06c463d7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-r2gbh" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240083 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/03021110-7732-4dff-a5e3-f481e8b7c0d6-service-ca\") pod \"console-f9d7485db-kk8mb\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240105 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac8c6176-0450-4050-9094-003067cd40b1-serving-cert\") pod \"openshift-config-operator-7777fb866f-kg4lw\" (UID: \"ac8c6176-0450-4050-9094-003067cd40b1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kg4lw" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240133 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6024dcd8-96d6-4c42-9670-42e3608cd7ee-client-ca\") pod \"route-controller-manager-6576b87f9c-tvqq9\" (UID: \"6024dcd8-96d6-4c42-9670-42e3608cd7ee\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240160 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a556e01d-7f37-46db-ac77-6a681cad9e7e-config\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240184 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20e7b7c5-1480-498e-ab4d-2e335da52dd3-config\") pod \"machine-approver-56656f9798-6jzdp\" (UID: \"20e7b7c5-1480-498e-ab4d-2e335da52dd3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jzdp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240208 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/235934fb-ec04-4784-b30c-5c80dfa325ad-serving-cert\") pod \"authentication-operator-69f744f599-5c2lp\" (UID: \"235934fb-ec04-4784-b30c-5c80dfa325ad\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5c2lp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240232 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm8zs\" (UniqueName: \"kubernetes.io/projected/7a975a4d-12df-4c73-b644-ff281b3c3390-kube-api-access-tm8zs\") pod \"cluster-image-registry-operator-dc59b4c8b-8lmkd\" (UID: \"7a975a4d-12df-4c73-b644-ff281b3c3390\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8lmkd" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240256 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rhgj\" (UniqueName: \"kubernetes.io/projected/c71a13b3-a727-4b88-9be8-954a62dc7c86-kube-api-access-4rhgj\") pod \"ingress-operator-5b745b69d9-8ft9r\" (UID: \"c71a13b3-a727-4b88-9be8-954a62dc7c86\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8ft9r" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240283 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03021110-7732-4dff-a5e3-f481e8b7c0d6-trusted-ca-bundle\") pod \"console-f9d7485db-kk8mb\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240306 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6024dcd8-96d6-4c42-9670-42e3608cd7ee-config\") pod \"route-controller-manager-6576b87f9c-tvqq9\" (UID: \"6024dcd8-96d6-4c42-9670-42e3608cd7ee\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240328 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6024dcd8-96d6-4c42-9670-42e3608cd7ee-serving-cert\") pod \"route-controller-manager-6576b87f9c-tvqq9\" (UID: \"6024dcd8-96d6-4c42-9670-42e3608cd7ee\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240352 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a556e01d-7f37-46db-ac77-6a681cad9e7e-etcd-serving-ca\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240380 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-542jg\" (UniqueName: \"kubernetes.io/projected/b46224fb-c151-4ba9-87c7-85e0ed6e11bc-kube-api-access-542jg\") pod \"apiserver-7bbb656c7d-bbss6\" (UID: \"b46224fb-c151-4ba9-87c7-85e0ed6e11bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240404 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/94922a63-40d9-4944-9d95-f860967ea5b3-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-9cdmt\" (UID: \"94922a63-40d9-4944-9d95-f860967ea5b3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9cdmt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240429 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7kj8\" (UniqueName: \"kubernetes.io/projected/2e58b61f-82bf-42c2-a664-cd3fcbd0fc03-kube-api-access-l7kj8\") pod \"downloads-7954f5f757-ppw5v\" (UID: \"2e58b61f-82bf-42c2-a664-cd3fcbd0fc03\") " pod="openshift-console/downloads-7954f5f757-ppw5v" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240452 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240492 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b46224fb-c151-4ba9-87c7-85e0ed6e11bc-audit-dir\") pod \"apiserver-7bbb656c7d-bbss6\" (UID: \"b46224fb-c151-4ba9-87c7-85e0ed6e11bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240516 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c2ff7dad-a4da-4243-ab1e-21b47230504d-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2kprn\" (UID: \"c2ff7dad-a4da-4243-ab1e-21b47230504d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240541 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/84e61bdd-8de2-4cd1-93a2-b43d802fbea3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-ctpbx\" (UID: \"84e61bdd-8de2-4cd1-93a2-b43d802fbea3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctpbx" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240569 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240597 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a556e01d-7f37-46db-ac77-6a681cad9e7e-node-pullsecrets\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240623 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptcqb\" (UniqueName: \"kubernetes.io/projected/ac8c6176-0450-4050-9094-003067cd40b1-kube-api-access-ptcqb\") pod \"openshift-config-operator-7777fb866f-kg4lw\" (UID: \"ac8c6176-0450-4050-9094-003067cd40b1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kg4lw" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240650 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240672 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/84e61bdd-8de2-4cd1-93a2-b43d802fbea3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-ctpbx\" (UID: \"84e61bdd-8de2-4cd1-93a2-b43d802fbea3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctpbx" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.240695 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/235934fb-ec04-4784-b30c-5c80dfa325ad-config\") pod \"authentication-operator-69f744f599-5c2lp\" (UID: \"235934fb-ec04-4784-b30c-5c80dfa325ad\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5c2lp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.241091 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.241459 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wfm8s"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.243195 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jjwjb"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.243701 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-k92zl"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.244019 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wfm8s" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.244212 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-k92zl" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.244300 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jjwjb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.245747 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/a556e01d-7f37-46db-ac77-6a681cad9e7e-image-import-ca\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.246299 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a556e01d-7f37-46db-ac77-6a681cad9e7e-config\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.247045 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b46224fb-c151-4ba9-87c7-85e0ed6e11bc-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-bbss6\" (UID: \"b46224fb-c151-4ba9-87c7-85e0ed6e11bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.247138 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.247976 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.248083 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a556e01d-7f37-46db-ac77-6a681cad9e7e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.248109 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.248220 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.248363 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.248448 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.248466 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.249332 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a556e01d-7f37-46db-ac77-6a681cad9e7e-audit-dir\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.249587 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2ff7dad-a4da-4243-ab1e-21b47230504d-config\") pod \"controller-manager-879f6c89f-2kprn\" (UID: \"c2ff7dad-a4da-4243-ab1e-21b47230504d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.250940 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/a556e01d-7f37-46db-ac77-6a681cad9e7e-audit\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.251027 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a556e01d-7f37-46db-ac77-6a681cad9e7e-node-pullsecrets\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.251110 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a556e01d-7f37-46db-ac77-6a681cad9e7e-etcd-serving-ca\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.251247 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b46224fb-c151-4ba9-87c7-85e0ed6e11bc-audit-dir\") pod \"apiserver-7bbb656c7d-bbss6\" (UID: \"b46224fb-c151-4ba9-87c7-85e0ed6e11bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.251706 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fnxt7"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.252196 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-fnxt7" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.254434 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/20e7b7c5-1480-498e-ab4d-2e335da52dd3-auth-proxy-config\") pod \"machine-approver-56656f9798-6jzdp\" (UID: \"20e7b7c5-1480-498e-ab4d-2e335da52dd3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jzdp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.256581 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b46224fb-c151-4ba9-87c7-85e0ed6e11bc-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-bbss6\" (UID: \"b46224fb-c151-4ba9-87c7-85e0ed6e11bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.256639 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c2ff7dad-a4da-4243-ab1e-21b47230504d-client-ca\") pod \"controller-manager-879f6c89f-2kprn\" (UID: \"c2ff7dad-a4da-4243-ab1e-21b47230504d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.256784 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b46224fb-c151-4ba9-87c7-85e0ed6e11bc-audit-policies\") pod \"apiserver-7bbb656c7d-bbss6\" (UID: \"b46224fb-c151-4ba9-87c7-85e0ed6e11bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.257052 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c2ff7dad-a4da-4243-ab1e-21b47230504d-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2kprn\" (UID: \"c2ff7dad-a4da-4243-ab1e-21b47230504d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.258705 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b46224fb-c151-4ba9-87c7-85e0ed6e11bc-encryption-config\") pod \"apiserver-7bbb656c7d-bbss6\" (UID: \"b46224fb-c151-4ba9-87c7-85e0ed6e11bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.258925 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b46224fb-c151-4ba9-87c7-85e0ed6e11bc-serving-cert\") pod \"apiserver-7bbb656c7d-bbss6\" (UID: \"b46224fb-c151-4ba9-87c7-85e0ed6e11bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.258951 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b46224fb-c151-4ba9-87c7-85e0ed6e11bc-etcd-client\") pod \"apiserver-7bbb656c7d-bbss6\" (UID: \"b46224fb-c151-4ba9-87c7-85e0ed6e11bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.259089 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-bhj2p"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.259483 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2ff7dad-a4da-4243-ab1e-21b47230504d-serving-cert\") pod \"controller-manager-879f6c89f-2kprn\" (UID: \"c2ff7dad-a4da-4243-ab1e-21b47230504d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.259748 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/94922a63-40d9-4944-9d95-f860967ea5b3-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-9cdmt\" (UID: \"94922a63-40d9-4944-9d95-f860967ea5b3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9cdmt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.259831 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.259986 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a556e01d-7f37-46db-ac77-6a681cad9e7e-serving-cert\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.260084 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a556e01d-7f37-46db-ac77-6a681cad9e7e-etcd-client\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.260139 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a556e01d-7f37-46db-ac77-6a681cad9e7e-encryption-config\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.260785 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bhj2p" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.261015 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l9b5"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.261150 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20e7b7c5-1480-498e-ab4d-2e335da52dd3-config\") pod \"machine-approver-56656f9798-6jzdp\" (UID: \"20e7b7c5-1480-498e-ab4d-2e335da52dd3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jzdp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.262773 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l9b5" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.265675 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zh6tw"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.266834 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2kprn"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.266955 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zh6tw" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.267240 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-v8c75"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.268076 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-v8c75" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.268560 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412960-6vr8n"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.270701 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-6vr8n" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.270872 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/20e7b7c5-1480-498e-ab4d-2e335da52dd3-machine-approver-tls\") pod \"machine-approver-56656f9798-6jzdp\" (UID: \"20e7b7c5-1480-498e-ab4d-2e335da52dd3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jzdp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.271961 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vns7l"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.272650 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.273253 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sn2r4"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.273906 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sn2r4" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.274143 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jmdsz"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.274688 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jmdsz" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.275178 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.275313 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hwftp"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.276040 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.276972 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.277835 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-5mthf"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.278771 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-54vrz"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.278785 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-5mthf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.279703 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-ggw6k"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.280607 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xd7m2"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.281495 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fnxt7"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.282429 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-8ft9r"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.283381 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nsq9q"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.284322 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qvmlb"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.285217 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wk6qg"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.286141 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-5c2lp"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.287146 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9cdmt"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.288103 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-4rq7l"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.289065 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5wc9c"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.289993 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jjwjb"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.290941 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rrj4k"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.291919 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-kg4lw"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.293743 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l9b5"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.294661 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.294781 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zh6tw"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.295702 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-k92zl"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.296638 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-b5hpf"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.297587 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wfm8s"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.298725 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-ppw5v"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.299564 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-b9f5r"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.300272 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-b9f5r" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.300552 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8lmkd"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.301598 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-dc55h"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.302981 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-v8c75"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.304541 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-kk8mb"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.306066 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-bhj2p"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.307315 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-r2gbh"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.308309 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-wsmkz"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.309302 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-wsmkz" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.309800 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vns7l"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.315509 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.315928 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sn2r4"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.318889 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jmdsz"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.320630 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412960-6vr8n"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.322222 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-55d4c"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.323582 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-5mthf"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.324778 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-b9f5r"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.325896 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-gbvbf"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.326396 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-gbvbf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.327350 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-wsmkz"] Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.341191 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6024dcd8-96d6-4c42-9670-42e3608cd7ee-client-ca\") pod \"route-controller-manager-6576b87f9c-tvqq9\" (UID: \"6024dcd8-96d6-4c42-9670-42e3608cd7ee\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.341238 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w5qv\" (UniqueName: \"kubernetes.io/projected/8c88814f-7a99-4878-9517-55cbecd72ff0-kube-api-access-9w5qv\") pod \"machine-config-controller-84d6567774-jjwjb\" (UID: \"8c88814f-7a99-4878-9517-55cbecd72ff0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jjwjb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.341267 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rhgj\" (UniqueName: \"kubernetes.io/projected/c71a13b3-a727-4b88-9be8-954a62dc7c86-kube-api-access-4rhgj\") pod \"ingress-operator-5b745b69d9-8ft9r\" (UID: \"c71a13b3-a727-4b88-9be8-954a62dc7c86\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8ft9r" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.341292 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03021110-7732-4dff-a5e3-f481e8b7c0d6-trusted-ca-bundle\") pod \"console-f9d7485db-kk8mb\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.341318 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm8zs\" (UniqueName: \"kubernetes.io/projected/7a975a4d-12df-4c73-b644-ff281b3c3390-kube-api-access-tm8zs\") pod \"cluster-image-registry-operator-dc59b4c8b-8lmkd\" (UID: \"7a975a4d-12df-4c73-b644-ff281b3c3390\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8lmkd" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.341343 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/75498a9e-2bcc-4e76-a528-16df0f6d8ae7-signing-cabundle\") pod \"service-ca-9c57cc56f-v8c75\" (UID: \"75498a9e-2bcc-4e76-a528-16df0f6d8ae7\") " pod="openshift-service-ca/service-ca-9c57cc56f-v8c75" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.341366 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6ce38b01-0835-4a45-a262-f584fe51bd62-cert\") pod \"ingress-canary-b9f5r\" (UID: \"6ce38b01-0835-4a45-a262-f584fe51bd62\") " pod="openshift-ingress-canary/ingress-canary-b9f5r" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.341385 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8c88814f-7a99-4878-9517-55cbecd72ff0-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jjwjb\" (UID: \"8c88814f-7a99-4878-9517-55cbecd72ff0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jjwjb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.341410 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6024dcd8-96d6-4c42-9670-42e3608cd7ee-config\") pod \"route-controller-manager-6576b87f9c-tvqq9\" (UID: \"6024dcd8-96d6-4c42-9670-42e3608cd7ee\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.341425 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9a6a4ae3-f219-4a1b-bfbe-f9f829945a8a-profile-collector-cert\") pod \"catalog-operator-68c6474976-xd7m2\" (UID: \"9a6a4ae3-f219-4a1b-bfbe-f9f829945a8a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xd7m2" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.341446 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/84e61bdd-8de2-4cd1-93a2-b43d802fbea3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-ctpbx\" (UID: \"84e61bdd-8de2-4cd1-93a2-b43d802fbea3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctpbx" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.341466 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7274d4ec-5476-46c7-ab6a-f4ed0f4b294e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jmdsz\" (UID: \"7274d4ec-5476-46c7-ab6a-f4ed0f4b294e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jmdsz" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.341515 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f089be95-2900-436d-b1f4-627095e2d563-images\") pod \"machine-config-operator-74547568cd-dc55h\" (UID: \"f089be95-2900-436d-b1f4-627095e2d563\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dc55h" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.341538 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/235934fb-ec04-4784-b30c-5c80dfa325ad-config\") pod \"authentication-operator-69f744f599-5c2lp\" (UID: \"235934fb-ec04-4784-b30c-5c80dfa325ad\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5c2lp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.341606 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/03021110-7732-4dff-a5e3-f481e8b7c0d6-console-serving-cert\") pod \"console-f9d7485db-kk8mb\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.341636 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2r7z\" (UniqueName: \"kubernetes.io/projected/56ad3380-4861-49d1-8758-7b1e27f74560-kube-api-access-n2r7z\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.342317 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbj97\" (UniqueName: \"kubernetes.io/projected/a6e6bbd8-c854-4708-b9b1-40884d1be6c7-kube-api-access-lbj97\") pod \"collect-profiles-29412960-6vr8n\" (UID: \"a6e6bbd8-c854-4708-b9b1-40884d1be6c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-6vr8n" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.342376 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a741e843-b85d-4250-844a-b03534dc47d2-metrics-tls\") pod \"dns-default-wsmkz\" (UID: \"a741e843-b85d-4250-844a-b03534dc47d2\") " pod="openshift-dns/dns-default-wsmkz" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.342408 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09d80f4a-297b-4000-8fcd-c586b5267339-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-nsq9q\" (UID: \"09d80f4a-297b-4000-8fcd-c586b5267339\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nsq9q" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.342431 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a29a797b-0ffe-4e7c-b4c4-d3869fb694fe-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rrj4k\" (UID: \"a29a797b-0ffe-4e7c-b4c4-d3869fb694fe\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rrj4k" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.342474 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xjhl\" (UniqueName: \"kubernetes.io/projected/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-kube-api-access-9xjhl\") pod \"etcd-operator-b45778765-vns7l\" (UID: \"77eb0bc5-1a4a-48ae-98cc-f26a450af04a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.342504 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/56ad3380-4861-49d1-8758-7b1e27f74560-audit-policies\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.342525 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e7a61bf3-1760-4910-9561-000d68ca9708-registration-dir\") pod \"csi-hostpathplugin-5mthf\" (UID: \"e7a61bf3-1760-4910-9561-000d68ca9708\") " pod="hostpath-provisioner/csi-hostpathplugin-5mthf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.342502 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/84e61bdd-8de2-4cd1-93a2-b43d802fbea3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-ctpbx\" (UID: \"84e61bdd-8de2-4cd1-93a2-b43d802fbea3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctpbx" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.342606 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.342667 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e1ff9891-7017-41a2-b7c0-183caa783671-trusted-ca\") pod \"console-operator-58897d9998-qvmlb\" (UID: \"e1ff9891-7017-41a2-b7c0-183caa783671\") " pod="openshift-console-operator/console-operator-58897d9998-qvmlb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.342704 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e7a61bf3-1760-4910-9561-000d68ca9708-plugins-dir\") pod \"csi-hostpathplugin-5mthf\" (UID: \"e7a61bf3-1760-4910-9561-000d68ca9708\") " pod="hostpath-provisioner/csi-hostpathplugin-5mthf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.342728 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/75498a9e-2bcc-4e76-a528-16df0f6d8ae7-signing-key\") pod \"service-ca-9c57cc56f-v8c75\" (UID: \"75498a9e-2bcc-4e76-a528-16df0f6d8ae7\") " pod="openshift-service-ca/service-ca-9c57cc56f-v8c75" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.342747 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/817c8b8a-cab9-474a-ac82-30b55677c2f4-config\") pod \"service-ca-operator-777779d784-k92zl\" (UID: \"817c8b8a-cab9-474a-ac82-30b55677c2f4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k92zl" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.342790 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e569f94d-9231-4277-af77-f97c23fcbabe-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-wk6qg\" (UID: \"e569f94d-9231-4277-af77-f97c23fcbabe\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wk6qg" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.342810 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-etcd-client\") pod \"etcd-operator-b45778765-vns7l\" (UID: \"77eb0bc5-1a4a-48ae-98cc-f26a450af04a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.342842 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/22c37be4-5e07-4638-aa01-04ee06c463d7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-r2gbh\" (UID: \"22c37be4-5e07-4638-aa01-04ee06c463d7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-r2gbh" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.342864 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-serving-cert\") pod \"etcd-operator-b45778765-vns7l\" (UID: \"77eb0bc5-1a4a-48ae-98cc-f26a450af04a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.342939 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/235934fb-ec04-4784-b30c-5c80dfa325ad-config\") pod \"authentication-operator-69f744f599-5c2lp\" (UID: \"235934fb-ec04-4784-b30c-5c80dfa325ad\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5c2lp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.343008 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.343063 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2dvn\" (UniqueName: \"kubernetes.io/projected/6ce38b01-0835-4a45-a262-f584fe51bd62-kube-api-access-c2dvn\") pod \"ingress-canary-b9f5r\" (UID: \"6ce38b01-0835-4a45-a262-f584fe51bd62\") " pod="openshift-ingress-canary/ingress-canary-b9f5r" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.343092 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e569f94d-9231-4277-af77-f97c23fcbabe-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-wk6qg\" (UID: \"e569f94d-9231-4277-af77-f97c23fcbabe\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wk6qg" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.343124 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6024dcd8-96d6-4c42-9670-42e3608cd7ee-config\") pod \"route-controller-manager-6576b87f9c-tvqq9\" (UID: \"6024dcd8-96d6-4c42-9670-42e3608cd7ee\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.343162 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a1b6dae-ff3d-4bfb-a811-98866c9c2b59-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-54vrz\" (UID: \"0a1b6dae-ff3d-4bfb-a811-98866c9c2b59\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-54vrz" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.343257 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6024dcd8-96d6-4c42-9670-42e3608cd7ee-client-ca\") pod \"route-controller-manager-6576b87f9c-tvqq9\" (UID: \"6024dcd8-96d6-4c42-9670-42e3608cd7ee\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.343335 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68597ef9-34cf-45c4-86c7-153f9f4d9363-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5wc9c\" (UID: \"68597ef9-34cf-45c4-86c7-153f9f4d9363\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5wc9c" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.343453 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-config\") pod \"etcd-operator-b45778765-vns7l\" (UID: \"77eb0bc5-1a4a-48ae-98cc-f26a450af04a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.343571 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c71a13b3-a727-4b88-9be8-954a62dc7c86-bound-sa-token\") pod \"ingress-operator-5b745b69d9-8ft9r\" (UID: \"c71a13b3-a727-4b88-9be8-954a62dc7c86\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8ft9r" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.343622 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.343716 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7a975a4d-12df-4c73-b644-ff281b3c3390-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8lmkd\" (UID: \"7a975a4d-12df-4c73-b644-ff281b3c3390\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8lmkd" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.344002 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c71a13b3-a727-4b88-9be8-954a62dc7c86-trusted-ca\") pod \"ingress-operator-5b745b69d9-8ft9r\" (UID: \"c71a13b3-a727-4b88-9be8-954a62dc7c86\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8ft9r" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.344064 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs7mz\" (UniqueName: \"kubernetes.io/projected/f089be95-2900-436d-b1f4-627095e2d563-kube-api-access-bs7mz\") pod \"machine-config-operator-74547568cd-dc55h\" (UID: \"f089be95-2900-436d-b1f4-627095e2d563\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dc55h" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.344101 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7lbl\" (UniqueName: \"kubernetes.io/projected/ef0bf68e-f7d5-43a2-890c-10ffba3e6238-kube-api-access-q7lbl\") pod \"migrator-59844c95c7-bhj2p\" (UID: \"ef0bf68e-f7d5-43a2-890c-10ffba3e6238\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bhj2p" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.344145 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxzlg\" (UniqueName: \"kubernetes.io/projected/31bfedfc-39de-4bf0-b049-e41c577ac848-kube-api-access-nxzlg\") pod \"olm-operator-6b444d44fb-wfm8s\" (UID: \"31bfedfc-39de-4bf0-b049-e41c577ac848\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wfm8s" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.344219 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjd2p\" (UniqueName: \"kubernetes.io/projected/03021110-7732-4dff-a5e3-f481e8b7c0d6-kube-api-access-pjd2p\") pod \"console-f9d7485db-kk8mb\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.344286 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/7a975a4d-12df-4c73-b644-ff281b3c3390-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8lmkd\" (UID: \"7a975a4d-12df-4c73-b644-ff281b3c3390\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8lmkd" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.344343 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a741e843-b85d-4250-844a-b03534dc47d2-config-volume\") pod \"dns-default-wsmkz\" (UID: \"a741e843-b85d-4250-844a-b03534dc47d2\") " pod="openshift-dns/dns-default-wsmkz" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.344443 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/817c8b8a-cab9-474a-ac82-30b55677c2f4-serving-cert\") pod \"service-ca-operator-777779d784-k92zl\" (UID: \"817c8b8a-cab9-474a-ac82-30b55677c2f4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k92zl" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.344483 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68597ef9-34cf-45c4-86c7-153f9f4d9363-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5wc9c\" (UID: \"68597ef9-34cf-45c4-86c7-153f9f4d9363\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5wc9c" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.344549 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zk89t\" (UniqueName: \"kubernetes.io/projected/892e64fc-db6b-4410-9b4c-3ff78cc82074-kube-api-access-zk89t\") pod \"machine-api-operator-5694c8668f-ggw6k\" (UID: \"892e64fc-db6b-4410-9b4c-3ff78cc82074\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ggw6k" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.344600 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/31bfedfc-39de-4bf0-b049-e41c577ac848-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wfm8s\" (UID: \"31bfedfc-39de-4bf0-b049-e41c577ac848\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wfm8s" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.344693 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/03021110-7732-4dff-a5e3-f481e8b7c0d6-console-config\") pod \"console-f9d7485db-kk8mb\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.344712 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1ff9891-7017-41a2-b7c0-183caa783671-config\") pod \"console-operator-58897d9998-qvmlb\" (UID: \"e1ff9891-7017-41a2-b7c0-183caa783671\") " pod="openshift-console-operator/console-operator-58897d9998-qvmlb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.344747 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a29a797b-0ffe-4e7c-b4c4-d3869fb694fe-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rrj4k\" (UID: \"a29a797b-0ffe-4e7c-b4c4-d3869fb694fe\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rrj4k" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.344824 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44wgw\" (UniqueName: \"kubernetes.io/projected/a741e843-b85d-4250-844a-b03534dc47d2-kube-api-access-44wgw\") pod \"dns-default-wsmkz\" (UID: \"a741e843-b85d-4250-844a-b03534dc47d2\") " pod="openshift-dns/dns-default-wsmkz" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.344848 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/ac8c6176-0450-4050-9094-003067cd40b1-available-featuregates\") pod \"openshift-config-operator-7777fb866f-kg4lw\" (UID: \"ac8c6176-0450-4050-9094-003067cd40b1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kg4lw" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.344893 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dvph\" (UniqueName: \"kubernetes.io/projected/12d866d2-1002-4db4-9832-59b5fea5062e-kube-api-access-7dvph\") pod \"control-plane-machine-set-operator-78cbb6b69f-sn2r4\" (UID: \"12d866d2-1002-4db4-9832-59b5fea5062e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sn2r4" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.344916 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-etcd-ca\") pod \"etcd-operator-b45778765-vns7l\" (UID: \"77eb0bc5-1a4a-48ae-98cc-f26a450af04a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.344938 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7a975a4d-12df-4c73-b644-ff281b3c3390-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8lmkd\" (UID: \"7a975a4d-12df-4c73-b644-ff281b3c3390\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8lmkd" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.344956 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/31bfedfc-39de-4bf0-b049-e41c577ac848-srv-cert\") pod \"olm-operator-6b444d44fb-wfm8s\" (UID: \"31bfedfc-39de-4bf0-b049-e41c577ac848\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wfm8s" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.344977 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a6e6bbd8-c854-4708-b9b1-40884d1be6c7-secret-volume\") pod \"collect-profiles-29412960-6vr8n\" (UID: \"a6e6bbd8-c854-4708-b9b1-40884d1be6c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-6vr8n" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345010 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84e61bdd-8de2-4cd1-93a2-b43d802fbea3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-ctpbx\" (UID: \"84e61bdd-8de2-4cd1-93a2-b43d802fbea3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctpbx" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345029 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e7a61bf3-1760-4910-9561-000d68ca9708-mountpoint-dir\") pod \"csi-hostpathplugin-5mthf\" (UID: \"e7a61bf3-1760-4910-9561-000d68ca9708\") " pod="hostpath-provisioner/csi-hostpathplugin-5mthf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345051 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frnjz\" (UniqueName: \"kubernetes.io/projected/9a6a4ae3-f219-4a1b-bfbe-f9f829945a8a-kube-api-access-frnjz\") pod \"catalog-operator-68c6474976-xd7m2\" (UID: \"9a6a4ae3-f219-4a1b-bfbe-f9f829945a8a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xd7m2" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345072 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/bff7045b-cac0-438c-b6a6-769147c3f108-tmpfs\") pod \"packageserver-d55dfcdfc-zh6tw\" (UID: \"bff7045b-cac0-438c-b6a6-769147c3f108\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zh6tw" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345044 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e1ff9891-7017-41a2-b7c0-183caa783671-trusted-ca\") pod \"console-operator-58897d9998-qvmlb\" (UID: \"e1ff9891-7017-41a2-b7c0-183caa783671\") " pod="openshift-console-operator/console-operator-58897d9998-qvmlb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345093 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-555gk\" (UniqueName: \"kubernetes.io/projected/cbae1035-ec59-4f24-884d-809c7e7bb685-kube-api-access-555gk\") pod \"router-default-5444994796-hnvvp\" (UID: \"cbae1035-ec59-4f24-884d-809c7e7bb685\") " pod="openshift-ingress/router-default-5444994796-hnvvp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345113 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/03021110-7732-4dff-a5e3-f481e8b7c0d6-service-ca\") pod \"console-f9d7485db-kk8mb\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345130 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac8c6176-0450-4050-9094-003067cd40b1-serving-cert\") pod \"openshift-config-operator-7777fb866f-kg4lw\" (UID: \"ac8c6176-0450-4050-9094-003067cd40b1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kg4lw" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345153 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zvj2\" (UniqueName: \"kubernetes.io/projected/bff7045b-cac0-438c-b6a6-769147c3f108-kube-api-access-4zvj2\") pod \"packageserver-d55dfcdfc-zh6tw\" (UID: \"bff7045b-cac0-438c-b6a6-769147c3f108\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zh6tw" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345179 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8c715a96-664c-4b26-8d5a-523a0fb4a17c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-fnxt7\" (UID: \"8c715a96-664c-4b26-8d5a-523a0fb4a17c\") " pod="openshift-marketplace/marketplace-operator-79b997595-fnxt7" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345198 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a6e6bbd8-c854-4708-b9b1-40884d1be6c7-config-volume\") pod \"collect-profiles-29412960-6vr8n\" (UID: \"a6e6bbd8-c854-4708-b9b1-40884d1be6c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-6vr8n" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345217 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/235934fb-ec04-4784-b30c-5c80dfa325ad-serving-cert\") pod \"authentication-operator-69f744f599-5c2lp\" (UID: \"235934fb-ec04-4784-b30c-5c80dfa325ad\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5c2lp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345239 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7kj8\" (UniqueName: \"kubernetes.io/projected/2e58b61f-82bf-42c2-a664-cd3fcbd0fc03-kube-api-access-l7kj8\") pod \"downloads-7954f5f757-ppw5v\" (UID: \"2e58b61f-82bf-42c2-a664-cd3fcbd0fc03\") " pod="openshift-console/downloads-7954f5f757-ppw5v" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345258 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6024dcd8-96d6-4c42-9670-42e3608cd7ee-serving-cert\") pod \"route-controller-manager-6576b87f9c-tvqq9\" (UID: \"6024dcd8-96d6-4c42-9670-42e3608cd7ee\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345297 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345319 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/cbae1035-ec59-4f24-884d-809c7e7bb685-default-certificate\") pod \"router-default-5444994796-hnvvp\" (UID: \"cbae1035-ec59-4f24-884d-809c7e7bb685\") " pod="openshift-ingress/router-default-5444994796-hnvvp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345345 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345365 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gt8xm\" (UniqueName: \"kubernetes.io/projected/8c715a96-664c-4b26-8d5a-523a0fb4a17c-kube-api-access-gt8xm\") pod \"marketplace-operator-79b997595-fnxt7\" (UID: \"8c715a96-664c-4b26-8d5a-523a0fb4a17c\") " pod="openshift-marketplace/marketplace-operator-79b997595-fnxt7" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345384 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qmnv\" (UniqueName: \"kubernetes.io/projected/e7a61bf3-1760-4910-9561-000d68ca9708-kube-api-access-6qmnv\") pod \"csi-hostpathplugin-5mthf\" (UID: \"e7a61bf3-1760-4910-9561-000d68ca9708\") " pod="hostpath-provisioner/csi-hostpathplugin-5mthf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345406 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptcqb\" (UniqueName: \"kubernetes.io/projected/ac8c6176-0450-4050-9094-003067cd40b1-kube-api-access-ptcqb\") pod \"openshift-config-operator-7777fb866f-kg4lw\" (UID: \"ac8c6176-0450-4050-9094-003067cd40b1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kg4lw" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345426 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c715a96-664c-4b26-8d5a-523a0fb4a17c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-fnxt7\" (UID: \"8c715a96-664c-4b26-8d5a-523a0fb4a17c\") " pod="openshift-marketplace/marketplace-operator-79b997595-fnxt7" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345445 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/84e61bdd-8de2-4cd1-93a2-b43d802fbea3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-ctpbx\" (UID: \"84e61bdd-8de2-4cd1-93a2-b43d802fbea3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctpbx" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345446 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/03021110-7732-4dff-a5e3-f481e8b7c0d6-console-serving-cert\") pod \"console-f9d7485db-kk8mb\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345466 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345487 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/12d866d2-1002-4db4-9832-59b5fea5062e-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-sn2r4\" (UID: \"12d866d2-1002-4db4-9832-59b5fea5062e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sn2r4" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345510 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a1b6dae-ff3d-4bfb-a811-98866c9c2b59-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-54vrz\" (UID: \"0a1b6dae-ff3d-4bfb-a811-98866c9c2b59\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-54vrz" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345573 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1ff9891-7017-41a2-b7c0-183caa783671-serving-cert\") pod \"console-operator-58897d9998-qvmlb\" (UID: \"e1ff9891-7017-41a2-b7c0-183caa783671\") " pod="openshift-console-operator/console-operator-58897d9998-qvmlb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345591 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/235934fb-ec04-4784-b30c-5c80dfa325ad-service-ca-bundle\") pod \"authentication-operator-69f744f599-5c2lp\" (UID: \"235934fb-ec04-4784-b30c-5c80dfa325ad\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5c2lp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345634 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02311a5a-a912-4cce-b076-d752266f0c95-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8l9b5\" (UID: \"02311a5a-a912-4cce-b076-d752266f0c95\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l9b5" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345656 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e7a61bf3-1760-4910-9561-000d68ca9708-socket-dir\") pod \"csi-hostpathplugin-5mthf\" (UID: \"e7a61bf3-1760-4910-9561-000d68ca9708\") " pod="hostpath-provisioner/csi-hostpathplugin-5mthf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345674 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bff7045b-cac0-438c-b6a6-769147c3f108-apiservice-cert\") pod \"packageserver-d55dfcdfc-zh6tw\" (UID: \"bff7045b-cac0-438c-b6a6-769147c3f108\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zh6tw" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345718 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/892e64fc-db6b-4410-9b4c-3ff78cc82074-images\") pod \"machine-api-operator-5694c8668f-ggw6k\" (UID: \"892e64fc-db6b-4410-9b4c-3ff78cc82074\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ggw6k" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345740 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02311a5a-a912-4cce-b076-d752266f0c95-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8l9b5\" (UID: \"02311a5a-a912-4cce-b076-d752266f0c95\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l9b5" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345792 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/84e61bdd-8de2-4cd1-93a2-b43d802fbea3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-ctpbx\" (UID: \"84e61bdd-8de2-4cd1-93a2-b43d802fbea3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctpbx" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345813 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bff7045b-cac0-438c-b6a6-769147c3f108-webhook-cert\") pod \"packageserver-d55dfcdfc-zh6tw\" (UID: \"bff7045b-cac0-438c-b6a6-769147c3f108\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zh6tw" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345858 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09d80f4a-297b-4000-8fcd-c586b5267339-config\") pod \"openshift-apiserver-operator-796bbdcf4f-nsq9q\" (UID: \"09d80f4a-297b-4000-8fcd-c586b5267339\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nsq9q" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345877 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cbae1035-ec59-4f24-884d-809c7e7bb685-metrics-certs\") pod \"router-default-5444994796-hnvvp\" (UID: \"cbae1035-ec59-4f24-884d-809c7e7bb685\") " pod="openshift-ingress/router-default-5444994796-hnvvp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345888 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/03021110-7732-4dff-a5e3-f481e8b7c0d6-console-config\") pod \"console-f9d7485db-kk8mb\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345901 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84e61bdd-8de2-4cd1-93a2-b43d802fbea3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-ctpbx\" (UID: \"84e61bdd-8de2-4cd1-93a2-b43d802fbea3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctpbx" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345979 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68597ef9-34cf-45c4-86c7-153f9f4d9363-config\") pod \"kube-controller-manager-operator-78b949d7b-5wc9c\" (UID: \"68597ef9-34cf-45c4-86c7-153f9f4d9363\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5wc9c" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.346017 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/57c54ccc-a846-4c00-8318-5d5331c70e70-metrics-tls\") pod \"dns-operator-744455d44c-4rq7l\" (UID: \"57c54ccc-a846-4c00-8318-5d5331c70e70\") " pod="openshift-dns-operator/dns-operator-744455d44c-4rq7l" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.346060 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8hgn\" (UniqueName: \"kubernetes.io/projected/57c54ccc-a846-4c00-8318-5d5331c70e70-kube-api-access-q8hgn\") pod \"dns-operator-744455d44c-4rq7l\" (UID: \"57c54ccc-a846-4c00-8318-5d5331c70e70\") " pod="openshift-dns-operator/dns-operator-744455d44c-4rq7l" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.346102 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzn2p\" (UniqueName: \"kubernetes.io/projected/6024dcd8-96d6-4c42-9670-42e3608cd7ee-kube-api-access-tzn2p\") pod \"route-controller-manager-6576b87f9c-tvqq9\" (UID: \"6024dcd8-96d6-4c42-9670-42e3608cd7ee\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.346161 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/cbae1035-ec59-4f24-884d-809c7e7bb685-stats-auth\") pod \"router-default-5444994796-hnvvp\" (UID: \"cbae1035-ec59-4f24-884d-809c7e7bb685\") " pod="openshift-ingress/router-default-5444994796-hnvvp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.346208 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8c88814f-7a99-4878-9517-55cbecd72ff0-proxy-tls\") pod \"machine-config-controller-84d6567774-jjwjb\" (UID: \"8c88814f-7a99-4878-9517-55cbecd72ff0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jjwjb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.346250 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.346303 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f089be95-2900-436d-b1f4-627095e2d563-auth-proxy-config\") pod \"machine-config-operator-74547568cd-dc55h\" (UID: \"f089be95-2900-436d-b1f4-627095e2d563\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dc55h" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.346346 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7r6j\" (UniqueName: \"kubernetes.io/projected/e1ff9891-7017-41a2-b7c0-183caa783671-kube-api-access-g7r6j\") pod \"console-operator-58897d9998-qvmlb\" (UID: \"e1ff9891-7017-41a2-b7c0-183caa783671\") " pod="openshift-console-operator/console-operator-58897d9998-qvmlb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.346422 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.346463 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.346508 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srczp\" (UniqueName: \"kubernetes.io/projected/7274d4ec-5476-46c7-ab6a-f4ed0f4b294e-kube-api-access-srczp\") pod \"package-server-manager-789f6589d5-jmdsz\" (UID: \"7274d4ec-5476-46c7-ab6a-f4ed0f4b294e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jmdsz" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.346552 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m985g\" (UniqueName: \"kubernetes.io/projected/817c8b8a-cab9-474a-ac82-30b55677c2f4-kube-api-access-m985g\") pod \"service-ca-operator-777779d784-k92zl\" (UID: \"817c8b8a-cab9-474a-ac82-30b55677c2f4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k92zl" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.346602 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqpn4\" (UniqueName: \"kubernetes.io/projected/09d80f4a-297b-4000-8fcd-c586b5267339-kube-api-access-gqpn4\") pod \"openshift-apiserver-operator-796bbdcf4f-nsq9q\" (UID: \"09d80f4a-297b-4000-8fcd-c586b5267339\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nsq9q" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.346659 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.346738 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.346784 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03021110-7732-4dff-a5e3-f481e8b7c0d6-trusted-ca-bundle\") pod \"console-f9d7485db-kk8mb\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.346805 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkxbv\" (UniqueName: \"kubernetes.io/projected/0a1b6dae-ff3d-4bfb-a811-98866c9c2b59-kube-api-access-hkxbv\") pod \"openshift-controller-manager-operator-756b6f6bc6-54vrz\" (UID: \"0a1b6dae-ff3d-4bfb-a811-98866c9c2b59\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-54vrz" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.346846 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1ff9891-7017-41a2-b7c0-183caa783671-config\") pod \"console-operator-58897d9998-qvmlb\" (UID: \"e1ff9891-7017-41a2-b7c0-183caa783671\") " pod="openshift-console-operator/console-operator-58897d9998-qvmlb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.346836 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/03021110-7732-4dff-a5e3-f481e8b7c0d6-oauth-serving-cert\") pod \"console-f9d7485db-kk8mb\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.346941 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e7a61bf3-1760-4910-9561-000d68ca9708-csi-data-dir\") pod \"csi-hostpathplugin-5mthf\" (UID: \"e7a61bf3-1760-4910-9561-000d68ca9708\") " pod="hostpath-provisioner/csi-hostpathplugin-5mthf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.346978 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-etcd-service-ca\") pod \"etcd-operator-b45778765-vns7l\" (UID: \"77eb0bc5-1a4a-48ae-98cc-f26a450af04a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.347021 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/56ad3380-4861-49d1-8758-7b1e27f74560-audit-dir\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.347054 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cbae1035-ec59-4f24-884d-809c7e7bb685-service-ca-bundle\") pod \"router-default-5444994796-hnvvp\" (UID: \"cbae1035-ec59-4f24-884d-809c7e7bb685\") " pod="openshift-ingress/router-default-5444994796-hnvvp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.347086 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9a6a4ae3-f219-4a1b-bfbe-f9f829945a8a-srv-cert\") pod \"catalog-operator-68c6474976-xd7m2\" (UID: \"9a6a4ae3-f219-4a1b-bfbe-f9f829945a8a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xd7m2" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.347121 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c71a13b3-a727-4b88-9be8-954a62dc7c86-metrics-tls\") pod \"ingress-operator-5b745b69d9-8ft9r\" (UID: \"c71a13b3-a727-4b88-9be8-954a62dc7c86\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8ft9r" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.347161 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/03021110-7732-4dff-a5e3-f481e8b7c0d6-console-oauth-config\") pod \"console-f9d7485db-kk8mb\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.347195 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nd5s5\" (UniqueName: \"kubernetes.io/projected/75498a9e-2bcc-4e76-a528-16df0f6d8ae7-kube-api-access-nd5s5\") pod \"service-ca-9c57cc56f-v8c75\" (UID: \"75498a9e-2bcc-4e76-a528-16df0f6d8ae7\") " pod="openshift-service-ca/service-ca-9c57cc56f-v8c75" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.347227 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9lqv\" (UniqueName: \"kubernetes.io/projected/02311a5a-a912-4cce-b076-d752266f0c95-kube-api-access-v9lqv\") pod \"kube-storage-version-migrator-operator-b67b599dd-8l9b5\" (UID: \"02311a5a-a912-4cce-b076-d752266f0c95\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l9b5" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.347265 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/892e64fc-db6b-4410-9b4c-3ff78cc82074-config\") pod \"machine-api-operator-5694c8668f-ggw6k\" (UID: \"892e64fc-db6b-4410-9b4c-3ff78cc82074\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ggw6k" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.347364 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e569f94d-9231-4277-af77-f97c23fcbabe-config\") pod \"kube-apiserver-operator-766d6c64bb-wk6qg\" (UID: \"e569f94d-9231-4277-af77-f97c23fcbabe\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wk6qg" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.347395 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/892e64fc-db6b-4410-9b4c-3ff78cc82074-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-ggw6k\" (UID: \"892e64fc-db6b-4410-9b4c-3ff78cc82074\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ggw6k" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.347417 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/235934fb-ec04-4784-b30c-5c80dfa325ad-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-5c2lp\" (UID: \"235934fb-ec04-4784-b30c-5c80dfa325ad\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5c2lp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.347443 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxtbv\" (UniqueName: \"kubernetes.io/projected/235934fb-ec04-4784-b30c-5c80dfa325ad-kube-api-access-bxtbv\") pod \"authentication-operator-69f744f599-5c2lp\" (UID: \"235934fb-ec04-4784-b30c-5c80dfa325ad\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5c2lp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.347466 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.347485 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a29a797b-0ffe-4e7c-b4c4-d3869fb694fe-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rrj4k\" (UID: \"a29a797b-0ffe-4e7c-b4c4-d3869fb694fe\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rrj4k" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.347511 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wst5d\" (UniqueName: \"kubernetes.io/projected/22c37be4-5e07-4638-aa01-04ee06c463d7-kube-api-access-wst5d\") pod \"multus-admission-controller-857f4d67dd-r2gbh\" (UID: \"22c37be4-5e07-4638-aa01-04ee06c463d7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-r2gbh" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.347530 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f089be95-2900-436d-b1f4-627095e2d563-proxy-tls\") pod \"machine-config-operator-74547568cd-dc55h\" (UID: \"f089be95-2900-436d-b1f4-627095e2d563\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dc55h" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.347793 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68597ef9-34cf-45c4-86c7-153f9f4d9363-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5wc9c\" (UID: \"68597ef9-34cf-45c4-86c7-153f9f4d9363\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5wc9c" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.347833 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e569f94d-9231-4277-af77-f97c23fcbabe-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-wk6qg\" (UID: \"e569f94d-9231-4277-af77-f97c23fcbabe\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wk6qg" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.345492 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/ac8c6176-0450-4050-9094-003067cd40b1-available-featuregates\") pod \"openshift-config-operator-7777fb866f-kg4lw\" (UID: \"ac8c6176-0450-4050-9094-003067cd40b1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kg4lw" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.347983 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c71a13b3-a727-4b88-9be8-954a62dc7c86-trusted-ca\") pod \"ingress-operator-5b745b69d9-8ft9r\" (UID: \"c71a13b3-a727-4b88-9be8-954a62dc7c86\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8ft9r" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.347987 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/03021110-7732-4dff-a5e3-f481e8b7c0d6-oauth-serving-cert\") pod \"console-f9d7485db-kk8mb\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.348118 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/84e61bdd-8de2-4cd1-93a2-b43d802fbea3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-ctpbx\" (UID: \"84e61bdd-8de2-4cd1-93a2-b43d802fbea3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctpbx" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.348143 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84e61bdd-8de2-4cd1-93a2-b43d802fbea3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-ctpbx\" (UID: \"84e61bdd-8de2-4cd1-93a2-b43d802fbea3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctpbx" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.348458 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7a975a4d-12df-4c73-b644-ff281b3c3390-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8lmkd\" (UID: \"7a975a4d-12df-4c73-b644-ff281b3c3390\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8lmkd" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.348543 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/03021110-7732-4dff-a5e3-f481e8b7c0d6-service-ca\") pod \"console-f9d7485db-kk8mb\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.348561 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/56ad3380-4861-49d1-8758-7b1e27f74560-audit-dir\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.348617 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/84e61bdd-8de2-4cd1-93a2-b43d802fbea3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-ctpbx\" (UID: \"84e61bdd-8de2-4cd1-93a2-b43d802fbea3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctpbx" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.348743 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a1b6dae-ff3d-4bfb-a811-98866c9c2b59-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-54vrz\" (UID: \"0a1b6dae-ff3d-4bfb-a811-98866c9c2b59\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-54vrz" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.348619 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/892e64fc-db6b-4410-9b4c-3ff78cc82074-images\") pod \"machine-api-operator-5694c8668f-ggw6k\" (UID: \"892e64fc-db6b-4410-9b4c-3ff78cc82074\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ggw6k" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.347369 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68597ef9-34cf-45c4-86c7-153f9f4d9363-config\") pod \"kube-controller-manager-operator-78b949d7b-5wc9c\" (UID: \"68597ef9-34cf-45c4-86c7-153f9f4d9363\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5wc9c" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.349082 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/235934fb-ec04-4784-b30c-5c80dfa325ad-service-ca-bundle\") pod \"authentication-operator-69f744f599-5c2lp\" (UID: \"235934fb-ec04-4784-b30c-5c80dfa325ad\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5c2lp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.349676 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e569f94d-9231-4277-af77-f97c23fcbabe-config\") pod \"kube-apiserver-operator-766d6c64bb-wk6qg\" (UID: \"e569f94d-9231-4277-af77-f97c23fcbabe\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wk6qg" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.349952 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09d80f4a-297b-4000-8fcd-c586b5267339-config\") pod \"openshift-apiserver-operator-796bbdcf4f-nsq9q\" (UID: \"09d80f4a-297b-4000-8fcd-c586b5267339\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nsq9q" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.350069 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6024dcd8-96d6-4c42-9670-42e3608cd7ee-serving-cert\") pod \"route-controller-manager-6576b87f9c-tvqq9\" (UID: \"6024dcd8-96d6-4c42-9670-42e3608cd7ee\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.350452 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/235934fb-ec04-4784-b30c-5c80dfa325ad-serving-cert\") pod \"authentication-operator-69f744f599-5c2lp\" (UID: \"235934fb-ec04-4784-b30c-5c80dfa325ad\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5c2lp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.350932 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/892e64fc-db6b-4410-9b4c-3ff78cc82074-config\") pod \"machine-api-operator-5694c8668f-ggw6k\" (UID: \"892e64fc-db6b-4410-9b4c-3ff78cc82074\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ggw6k" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.351216 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/235934fb-ec04-4784-b30c-5c80dfa325ad-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-5c2lp\" (UID: \"235934fb-ec04-4784-b30c-5c80dfa325ad\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5c2lp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.351491 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09d80f4a-297b-4000-8fcd-c586b5267339-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-nsq9q\" (UID: \"09d80f4a-297b-4000-8fcd-c586b5267339\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nsq9q" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.351507 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/892e64fc-db6b-4410-9b4c-3ff78cc82074-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-ggw6k\" (UID: \"892e64fc-db6b-4410-9b4c-3ff78cc82074\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ggw6k" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.351571 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/03021110-7732-4dff-a5e3-f481e8b7c0d6-console-oauth-config\") pod \"console-f9d7485db-kk8mb\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.351672 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac8c6176-0450-4050-9094-003067cd40b1-serving-cert\") pod \"openshift-config-operator-7777fb866f-kg4lw\" (UID: \"ac8c6176-0450-4050-9094-003067cd40b1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kg4lw" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.351939 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1ff9891-7017-41a2-b7c0-183caa783671-serving-cert\") pod \"console-operator-58897d9998-qvmlb\" (UID: \"e1ff9891-7017-41a2-b7c0-183caa783671\") " pod="openshift-console-operator/console-operator-58897d9998-qvmlb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.352114 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a1b6dae-ff3d-4bfb-a811-98866c9c2b59-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-54vrz\" (UID: \"0a1b6dae-ff3d-4bfb-a811-98866c9c2b59\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-54vrz" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.352744 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/57c54ccc-a846-4c00-8318-5d5331c70e70-metrics-tls\") pod \"dns-operator-744455d44c-4rq7l\" (UID: \"57c54ccc-a846-4c00-8318-5d5331c70e70\") " pod="openshift-dns-operator/dns-operator-744455d44c-4rq7l" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.353832 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c71a13b3-a727-4b88-9be8-954a62dc7c86-metrics-tls\") pod \"ingress-operator-5b745b69d9-8ft9r\" (UID: \"c71a13b3-a727-4b88-9be8-954a62dc7c86\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8ft9r" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.355485 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.367568 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/7a975a4d-12df-4c73-b644-ff281b3c3390-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8lmkd\" (UID: \"7a975a4d-12df-4c73-b644-ff281b3c3390\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8lmkd" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.375014 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.395390 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.408371 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/22c37be4-5e07-4638-aa01-04ee06c463d7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-r2gbh\" (UID: \"22c37be4-5e07-4638-aa01-04ee06c463d7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-r2gbh" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.430047 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.435322 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.439859 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.448424 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w5qv\" (UniqueName: \"kubernetes.io/projected/8c88814f-7a99-4878-9517-55cbecd72ff0-kube-api-access-9w5qv\") pod \"machine-config-controller-84d6567774-jjwjb\" (UID: \"8c88814f-7a99-4878-9517-55cbecd72ff0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jjwjb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.448479 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/75498a9e-2bcc-4e76-a528-16df0f6d8ae7-signing-cabundle\") pod \"service-ca-9c57cc56f-v8c75\" (UID: \"75498a9e-2bcc-4e76-a528-16df0f6d8ae7\") " pod="openshift-service-ca/service-ca-9c57cc56f-v8c75" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.448511 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8c88814f-7a99-4878-9517-55cbecd72ff0-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jjwjb\" (UID: \"8c88814f-7a99-4878-9517-55cbecd72ff0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jjwjb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.448533 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6ce38b01-0835-4a45-a262-f584fe51bd62-cert\") pod \"ingress-canary-b9f5r\" (UID: \"6ce38b01-0835-4a45-a262-f584fe51bd62\") " pod="openshift-ingress-canary/ingress-canary-b9f5r" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.448556 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9a6a4ae3-f219-4a1b-bfbe-f9f829945a8a-profile-collector-cert\") pod \"catalog-operator-68c6474976-xd7m2\" (UID: \"9a6a4ae3-f219-4a1b-bfbe-f9f829945a8a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xd7m2" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.448578 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7274d4ec-5476-46c7-ab6a-f4ed0f4b294e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jmdsz\" (UID: \"7274d4ec-5476-46c7-ab6a-f4ed0f4b294e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jmdsz" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.448602 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f089be95-2900-436d-b1f4-627095e2d563-images\") pod \"machine-config-operator-74547568cd-dc55h\" (UID: \"f089be95-2900-436d-b1f4-627095e2d563\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dc55h" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.448657 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbj97\" (UniqueName: \"kubernetes.io/projected/a6e6bbd8-c854-4708-b9b1-40884d1be6c7-kube-api-access-lbj97\") pod \"collect-profiles-29412960-6vr8n\" (UID: \"a6e6bbd8-c854-4708-b9b1-40884d1be6c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-6vr8n" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.448688 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a741e843-b85d-4250-844a-b03534dc47d2-metrics-tls\") pod \"dns-default-wsmkz\" (UID: \"a741e843-b85d-4250-844a-b03534dc47d2\") " pod="openshift-dns/dns-default-wsmkz" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.448724 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a29a797b-0ffe-4e7c-b4c4-d3869fb694fe-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rrj4k\" (UID: \"a29a797b-0ffe-4e7c-b4c4-d3869fb694fe\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rrj4k" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.448790 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xjhl\" (UniqueName: \"kubernetes.io/projected/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-kube-api-access-9xjhl\") pod \"etcd-operator-b45778765-vns7l\" (UID: \"77eb0bc5-1a4a-48ae-98cc-f26a450af04a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.448825 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e7a61bf3-1760-4910-9561-000d68ca9708-registration-dir\") pod \"csi-hostpathplugin-5mthf\" (UID: \"e7a61bf3-1760-4910-9561-000d68ca9708\") " pod="hostpath-provisioner/csi-hostpathplugin-5mthf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.448867 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e7a61bf3-1760-4910-9561-000d68ca9708-plugins-dir\") pod \"csi-hostpathplugin-5mthf\" (UID: \"e7a61bf3-1760-4910-9561-000d68ca9708\") " pod="hostpath-provisioner/csi-hostpathplugin-5mthf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.448890 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/75498a9e-2bcc-4e76-a528-16df0f6d8ae7-signing-key\") pod \"service-ca-9c57cc56f-v8c75\" (UID: \"75498a9e-2bcc-4e76-a528-16df0f6d8ae7\") " pod="openshift-service-ca/service-ca-9c57cc56f-v8c75" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.448911 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/817c8b8a-cab9-474a-ac82-30b55677c2f4-config\") pod \"service-ca-operator-777779d784-k92zl\" (UID: \"817c8b8a-cab9-474a-ac82-30b55677c2f4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k92zl" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.448942 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-etcd-client\") pod \"etcd-operator-b45778765-vns7l\" (UID: \"77eb0bc5-1a4a-48ae-98cc-f26a450af04a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.448971 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-serving-cert\") pod \"etcd-operator-b45778765-vns7l\" (UID: \"77eb0bc5-1a4a-48ae-98cc-f26a450af04a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.449002 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2dvn\" (UniqueName: \"kubernetes.io/projected/6ce38b01-0835-4a45-a262-f584fe51bd62-kube-api-access-c2dvn\") pod \"ingress-canary-b9f5r\" (UID: \"6ce38b01-0835-4a45-a262-f584fe51bd62\") " pod="openshift-ingress-canary/ingress-canary-b9f5r" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.449039 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-config\") pod \"etcd-operator-b45778765-vns7l\" (UID: \"77eb0bc5-1a4a-48ae-98cc-f26a450af04a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.449091 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7lbl\" (UniqueName: \"kubernetes.io/projected/ef0bf68e-f7d5-43a2-890c-10ffba3e6238-kube-api-access-q7lbl\") pod \"migrator-59844c95c7-bhj2p\" (UID: \"ef0bf68e-f7d5-43a2-890c-10ffba3e6238\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bhj2p" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.449120 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxzlg\" (UniqueName: \"kubernetes.io/projected/31bfedfc-39de-4bf0-b049-e41c577ac848-kube-api-access-nxzlg\") pod \"olm-operator-6b444d44fb-wfm8s\" (UID: \"31bfedfc-39de-4bf0-b049-e41c577ac848\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wfm8s" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.449148 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs7mz\" (UniqueName: \"kubernetes.io/projected/f089be95-2900-436d-b1f4-627095e2d563-kube-api-access-bs7mz\") pod \"machine-config-operator-74547568cd-dc55h\" (UID: \"f089be95-2900-436d-b1f4-627095e2d563\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dc55h" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.449173 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a741e843-b85d-4250-844a-b03534dc47d2-config-volume\") pod \"dns-default-wsmkz\" (UID: \"a741e843-b85d-4250-844a-b03534dc47d2\") " pod="openshift-dns/dns-default-wsmkz" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.449196 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/817c8b8a-cab9-474a-ac82-30b55677c2f4-serving-cert\") pod \"service-ca-operator-777779d784-k92zl\" (UID: \"817c8b8a-cab9-474a-ac82-30b55677c2f4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k92zl" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.449237 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/31bfedfc-39de-4bf0-b049-e41c577ac848-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wfm8s\" (UID: \"31bfedfc-39de-4bf0-b049-e41c577ac848\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wfm8s" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.449263 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a29a797b-0ffe-4e7c-b4c4-d3869fb694fe-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rrj4k\" (UID: \"a29a797b-0ffe-4e7c-b4c4-d3869fb694fe\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rrj4k" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.449290 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dvph\" (UniqueName: \"kubernetes.io/projected/12d866d2-1002-4db4-9832-59b5fea5062e-kube-api-access-7dvph\") pod \"control-plane-machine-set-operator-78cbb6b69f-sn2r4\" (UID: \"12d866d2-1002-4db4-9832-59b5fea5062e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sn2r4" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.449317 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44wgw\" (UniqueName: \"kubernetes.io/projected/a741e843-b85d-4250-844a-b03534dc47d2-kube-api-access-44wgw\") pod \"dns-default-wsmkz\" (UID: \"a741e843-b85d-4250-844a-b03534dc47d2\") " pod="openshift-dns/dns-default-wsmkz" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.449341 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-etcd-ca\") pod \"etcd-operator-b45778765-vns7l\" (UID: \"77eb0bc5-1a4a-48ae-98cc-f26a450af04a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.449351 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e7a61bf3-1760-4910-9561-000d68ca9708-plugins-dir\") pod \"csi-hostpathplugin-5mthf\" (UID: \"e7a61bf3-1760-4910-9561-000d68ca9708\") " pod="hostpath-provisioner/csi-hostpathplugin-5mthf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.449369 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/31bfedfc-39de-4bf0-b049-e41c577ac848-srv-cert\") pod \"olm-operator-6b444d44fb-wfm8s\" (UID: \"31bfedfc-39de-4bf0-b049-e41c577ac848\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wfm8s" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.449464 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a6e6bbd8-c854-4708-b9b1-40884d1be6c7-secret-volume\") pod \"collect-profiles-29412960-6vr8n\" (UID: \"a6e6bbd8-c854-4708-b9b1-40884d1be6c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-6vr8n" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.449568 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frnjz\" (UniqueName: \"kubernetes.io/projected/9a6a4ae3-f219-4a1b-bfbe-f9f829945a8a-kube-api-access-frnjz\") pod \"catalog-operator-68c6474976-xd7m2\" (UID: \"9a6a4ae3-f219-4a1b-bfbe-f9f829945a8a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xd7m2" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.449617 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/bff7045b-cac0-438c-b6a6-769147c3f108-tmpfs\") pod \"packageserver-d55dfcdfc-zh6tw\" (UID: \"bff7045b-cac0-438c-b6a6-769147c3f108\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zh6tw" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.449674 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-555gk\" (UniqueName: \"kubernetes.io/projected/cbae1035-ec59-4f24-884d-809c7e7bb685-kube-api-access-555gk\") pod \"router-default-5444994796-hnvvp\" (UID: \"cbae1035-ec59-4f24-884d-809c7e7bb685\") " pod="openshift-ingress/router-default-5444994796-hnvvp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.449708 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e7a61bf3-1760-4910-9561-000d68ca9708-mountpoint-dir\") pod \"csi-hostpathplugin-5mthf\" (UID: \"e7a61bf3-1760-4910-9561-000d68ca9708\") " pod="hostpath-provisioner/csi-hostpathplugin-5mthf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.449747 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zvj2\" (UniqueName: \"kubernetes.io/projected/bff7045b-cac0-438c-b6a6-769147c3f108-kube-api-access-4zvj2\") pod \"packageserver-d55dfcdfc-zh6tw\" (UID: \"bff7045b-cac0-438c-b6a6-769147c3f108\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zh6tw" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.449802 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e7a61bf3-1760-4910-9561-000d68ca9708-mountpoint-dir\") pod \"csi-hostpathplugin-5mthf\" (UID: \"e7a61bf3-1760-4910-9561-000d68ca9708\") " pod="hostpath-provisioner/csi-hostpathplugin-5mthf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.449822 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8c715a96-664c-4b26-8d5a-523a0fb4a17c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-fnxt7\" (UID: \"8c715a96-664c-4b26-8d5a-523a0fb4a17c\") " pod="openshift-marketplace/marketplace-operator-79b997595-fnxt7" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.449864 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a6e6bbd8-c854-4708-b9b1-40884d1be6c7-config-volume\") pod \"collect-profiles-29412960-6vr8n\" (UID: \"a6e6bbd8-c854-4708-b9b1-40884d1be6c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-6vr8n" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.449948 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/cbae1035-ec59-4f24-884d-809c7e7bb685-default-certificate\") pod \"router-default-5444994796-hnvvp\" (UID: \"cbae1035-ec59-4f24-884d-809c7e7bb685\") " pod="openshift-ingress/router-default-5444994796-hnvvp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.450015 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gt8xm\" (UniqueName: \"kubernetes.io/projected/8c715a96-664c-4b26-8d5a-523a0fb4a17c-kube-api-access-gt8xm\") pod \"marketplace-operator-79b997595-fnxt7\" (UID: \"8c715a96-664c-4b26-8d5a-523a0fb4a17c\") " pod="openshift-marketplace/marketplace-operator-79b997595-fnxt7" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.450047 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qmnv\" (UniqueName: \"kubernetes.io/projected/e7a61bf3-1760-4910-9561-000d68ca9708-kube-api-access-6qmnv\") pod \"csi-hostpathplugin-5mthf\" (UID: \"e7a61bf3-1760-4910-9561-000d68ca9708\") " pod="hostpath-provisioner/csi-hostpathplugin-5mthf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.450109 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c715a96-664c-4b26-8d5a-523a0fb4a17c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-fnxt7\" (UID: \"8c715a96-664c-4b26-8d5a-523a0fb4a17c\") " pod="openshift-marketplace/marketplace-operator-79b997595-fnxt7" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.450145 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/12d866d2-1002-4db4-9832-59b5fea5062e-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-sn2r4\" (UID: \"12d866d2-1002-4db4-9832-59b5fea5062e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sn2r4" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.450201 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02311a5a-a912-4cce-b076-d752266f0c95-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8l9b5\" (UID: \"02311a5a-a912-4cce-b076-d752266f0c95\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l9b5" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.450237 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bff7045b-cac0-438c-b6a6-769147c3f108-apiservice-cert\") pod \"packageserver-d55dfcdfc-zh6tw\" (UID: \"bff7045b-cac0-438c-b6a6-769147c3f108\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zh6tw" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.450270 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e7a61bf3-1760-4910-9561-000d68ca9708-socket-dir\") pod \"csi-hostpathplugin-5mthf\" (UID: \"e7a61bf3-1760-4910-9561-000d68ca9708\") " pod="hostpath-provisioner/csi-hostpathplugin-5mthf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.450288 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e7a61bf3-1760-4910-9561-000d68ca9708-registration-dir\") pod \"csi-hostpathplugin-5mthf\" (UID: \"e7a61bf3-1760-4910-9561-000d68ca9708\") " pod="hostpath-provisioner/csi-hostpathplugin-5mthf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.450326 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02311a5a-a912-4cce-b076-d752266f0c95-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8l9b5\" (UID: \"02311a5a-a912-4cce-b076-d752266f0c95\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l9b5" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.450376 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e7a61bf3-1760-4910-9561-000d68ca9708-socket-dir\") pod \"csi-hostpathplugin-5mthf\" (UID: \"e7a61bf3-1760-4910-9561-000d68ca9708\") " pod="hostpath-provisioner/csi-hostpathplugin-5mthf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.450407 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/bff7045b-cac0-438c-b6a6-769147c3f108-tmpfs\") pod \"packageserver-d55dfcdfc-zh6tw\" (UID: \"bff7045b-cac0-438c-b6a6-769147c3f108\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zh6tw" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.450415 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bff7045b-cac0-438c-b6a6-769147c3f108-webhook-cert\") pod \"packageserver-d55dfcdfc-zh6tw\" (UID: \"bff7045b-cac0-438c-b6a6-769147c3f108\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zh6tw" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.450496 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cbae1035-ec59-4f24-884d-809c7e7bb685-metrics-certs\") pod \"router-default-5444994796-hnvvp\" (UID: \"cbae1035-ec59-4f24-884d-809c7e7bb685\") " pod="openshift-ingress/router-default-5444994796-hnvvp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.450587 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/cbae1035-ec59-4f24-884d-809c7e7bb685-stats-auth\") pod \"router-default-5444994796-hnvvp\" (UID: \"cbae1035-ec59-4f24-884d-809c7e7bb685\") " pod="openshift-ingress/router-default-5444994796-hnvvp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.450668 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8c88814f-7a99-4878-9517-55cbecd72ff0-proxy-tls\") pod \"machine-config-controller-84d6567774-jjwjb\" (UID: \"8c88814f-7a99-4878-9517-55cbecd72ff0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jjwjb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.450782 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f089be95-2900-436d-b1f4-627095e2d563-auth-proxy-config\") pod \"machine-config-operator-74547568cd-dc55h\" (UID: \"f089be95-2900-436d-b1f4-627095e2d563\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dc55h" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.450829 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srczp\" (UniqueName: \"kubernetes.io/projected/7274d4ec-5476-46c7-ab6a-f4ed0f4b294e-kube-api-access-srczp\") pod \"package-server-manager-789f6589d5-jmdsz\" (UID: \"7274d4ec-5476-46c7-ab6a-f4ed0f4b294e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jmdsz" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.450879 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8c88814f-7a99-4878-9517-55cbecd72ff0-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-jjwjb\" (UID: \"8c88814f-7a99-4878-9517-55cbecd72ff0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jjwjb" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.450883 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m985g\" (UniqueName: \"kubernetes.io/projected/817c8b8a-cab9-474a-ac82-30b55677c2f4-kube-api-access-m985g\") pod \"service-ca-operator-777779d784-k92zl\" (UID: \"817c8b8a-cab9-474a-ac82-30b55677c2f4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k92zl" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.451015 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e7a61bf3-1760-4910-9561-000d68ca9708-csi-data-dir\") pod \"csi-hostpathplugin-5mthf\" (UID: \"e7a61bf3-1760-4910-9561-000d68ca9708\") " pod="hostpath-provisioner/csi-hostpathplugin-5mthf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.451045 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-etcd-service-ca\") pod \"etcd-operator-b45778765-vns7l\" (UID: \"77eb0bc5-1a4a-48ae-98cc-f26a450af04a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.451075 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cbae1035-ec59-4f24-884d-809c7e7bb685-service-ca-bundle\") pod \"router-default-5444994796-hnvvp\" (UID: \"cbae1035-ec59-4f24-884d-809c7e7bb685\") " pod="openshift-ingress/router-default-5444994796-hnvvp" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.451100 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9a6a4ae3-f219-4a1b-bfbe-f9f829945a8a-srv-cert\") pod \"catalog-operator-68c6474976-xd7m2\" (UID: \"9a6a4ae3-f219-4a1b-bfbe-f9f829945a8a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xd7m2" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.451125 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nd5s5\" (UniqueName: \"kubernetes.io/projected/75498a9e-2bcc-4e76-a528-16df0f6d8ae7-kube-api-access-nd5s5\") pod \"service-ca-9c57cc56f-v8c75\" (UID: \"75498a9e-2bcc-4e76-a528-16df0f6d8ae7\") " pod="openshift-service-ca/service-ca-9c57cc56f-v8c75" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.451181 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9lqv\" (UniqueName: \"kubernetes.io/projected/02311a5a-a912-4cce-b076-d752266f0c95-kube-api-access-v9lqv\") pod \"kube-storage-version-migrator-operator-b67b599dd-8l9b5\" (UID: \"02311a5a-a912-4cce-b076-d752266f0c95\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l9b5" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.451225 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a29a797b-0ffe-4e7c-b4c4-d3869fb694fe-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rrj4k\" (UID: \"a29a797b-0ffe-4e7c-b4c4-d3869fb694fe\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rrj4k" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.451257 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f089be95-2900-436d-b1f4-627095e2d563-proxy-tls\") pod \"machine-config-operator-74547568cd-dc55h\" (UID: \"f089be95-2900-436d-b1f4-627095e2d563\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dc55h" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.451362 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e7a61bf3-1760-4910-9561-000d68ca9708-csi-data-dir\") pod \"csi-hostpathplugin-5mthf\" (UID: \"e7a61bf3-1760-4910-9561-000d68ca9708\") " pod="hostpath-provisioner/csi-hostpathplugin-5mthf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.451883 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f089be95-2900-436d-b1f4-627095e2d563-auth-proxy-config\") pod \"machine-config-operator-74547568cd-dc55h\" (UID: \"f089be95-2900-436d-b1f4-627095e2d563\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dc55h" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.456297 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.466484 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.474907 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.483512 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.495432 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.505919 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.528783 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.537981 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.541801 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.550626 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.555578 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.560820 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.574750 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.595626 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.603589 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.616127 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.623121 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.635553 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.655389 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.674342 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.677651 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.677653 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.677678 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.684054 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/56ad3380-4861-49d1-8758-7b1e27f74560-audit-policies\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.695707 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.705551 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.716048 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.734951 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.739446 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.755858 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.776828 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.783595 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a29a797b-0ffe-4e7c-b4c4-d3869fb694fe-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rrj4k\" (UID: \"a29a797b-0ffe-4e7c-b4c4-d3869fb694fe\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rrj4k" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.796177 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.802888 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a29a797b-0ffe-4e7c-b4c4-d3869fb694fe-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rrj4k\" (UID: \"a29a797b-0ffe-4e7c-b4c4-d3869fb694fe\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rrj4k" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.835567 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.856153 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.876368 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.896036 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.916165 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.935509 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.946944 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9a6a4ae3-f219-4a1b-bfbe-f9f829945a8a-srv-cert\") pod \"catalog-operator-68c6474976-xd7m2\" (UID: \"9a6a4ae3-f219-4a1b-bfbe-f9f829945a8a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xd7m2" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.955908 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.965533 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/31bfedfc-39de-4bf0-b049-e41c577ac848-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wfm8s\" (UID: \"31bfedfc-39de-4bf0-b049-e41c577ac848\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wfm8s" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.965558 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a6e6bbd8-c854-4708-b9b1-40884d1be6c7-secret-volume\") pod \"collect-profiles-29412960-6vr8n\" (UID: \"a6e6bbd8-c854-4708-b9b1-40884d1be6c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-6vr8n" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.967488 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9a6a4ae3-f219-4a1b-bfbe-f9f829945a8a-profile-collector-cert\") pod \"catalog-operator-68c6474976-xd7m2\" (UID: \"9a6a4ae3-f219-4a1b-bfbe-f9f829945a8a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xd7m2" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.976188 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 16:05:08 crc kubenswrapper[4998]: I1203 16:05:08.995918 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.015243 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.034962 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.044807 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f089be95-2900-436d-b1f4-627095e2d563-proxy-tls\") pod \"machine-config-operator-74547568cd-dc55h\" (UID: \"f089be95-2900-436d-b1f4-627095e2d563\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dc55h" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.055691 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.061072 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f089be95-2900-436d-b1f4-627095e2d563-images\") pod \"machine-config-operator-74547568cd-dc55h\" (UID: \"f089be95-2900-436d-b1f4-627095e2d563\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dc55h" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.075598 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.095227 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.105062 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/cbae1035-ec59-4f24-884d-809c7e7bb685-default-certificate\") pod \"router-default-5444994796-hnvvp\" (UID: \"cbae1035-ec59-4f24-884d-809c7e7bb685\") " pod="openshift-ingress/router-default-5444994796-hnvvp" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.115540 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.131529 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/cbae1035-ec59-4f24-884d-809c7e7bb685-stats-auth\") pod \"router-default-5444994796-hnvvp\" (UID: \"cbae1035-ec59-4f24-884d-809c7e7bb685\") " pod="openshift-ingress/router-default-5444994796-hnvvp" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.136137 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.145740 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cbae1035-ec59-4f24-884d-809c7e7bb685-metrics-certs\") pod \"router-default-5444994796-hnvvp\" (UID: \"cbae1035-ec59-4f24-884d-809c7e7bb685\") " pod="openshift-ingress/router-default-5444994796-hnvvp" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.155220 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.162953 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cbae1035-ec59-4f24-884d-809c7e7bb685-service-ca-bundle\") pod \"router-default-5444994796-hnvvp\" (UID: \"cbae1035-ec59-4f24-884d-809c7e7bb685\") " pod="openshift-ingress/router-default-5444994796-hnvvp" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.176141 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.216350 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.219450 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-542jg\" (UniqueName: \"kubernetes.io/projected/b46224fb-c151-4ba9-87c7-85e0ed6e11bc-kube-api-access-542jg\") pod \"apiserver-7bbb656c7d-bbss6\" (UID: \"b46224fb-c151-4ba9-87c7-85e0ed6e11bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.223935 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/31bfedfc-39de-4bf0-b049-e41c577ac848-srv-cert\") pod \"olm-operator-6b444d44fb-wfm8s\" (UID: \"31bfedfc-39de-4bf0-b049-e41c577ac848\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wfm8s" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.235065 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.244348 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8c88814f-7a99-4878-9517-55cbecd72ff0-proxy-tls\") pod \"machine-config-controller-84d6567774-jjwjb\" (UID: \"8c88814f-7a99-4878-9517-55cbecd72ff0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jjwjb" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.254342 4998 request.go:700] Waited for 1.009860598s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmachine-config-controller-dockercfg-c2lfx&limit=500&resourceVersion=0 Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.256402 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.276031 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.296547 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.315447 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.319955 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/817c8b8a-cab9-474a-ac82-30b55677c2f4-config\") pod \"service-ca-operator-777779d784-k92zl\" (UID: \"817c8b8a-cab9-474a-ac82-30b55677c2f4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k92zl" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.334634 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.343387 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/817c8b8a-cab9-474a-ac82-30b55677c2f4-serving-cert\") pod \"service-ca-operator-777779d784-k92zl\" (UID: \"817c8b8a-cab9-474a-ac82-30b55677c2f4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k92zl" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.355169 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.376470 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.393987 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbnql\" (UniqueName: \"kubernetes.io/projected/a556e01d-7f37-46db-ac77-6a681cad9e7e-kube-api-access-sbnql\") pod \"apiserver-76f77b778f-hwftp\" (UID: \"a556e01d-7f37-46db-ac77-6a681cad9e7e\") " pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.418470 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxdjj\" (UniqueName: \"kubernetes.io/projected/c2ff7dad-a4da-4243-ab1e-21b47230504d-kube-api-access-bxdjj\") pod \"controller-manager-879f6c89f-2kprn\" (UID: \"c2ff7dad-a4da-4243-ab1e-21b47230504d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.419506 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.435473 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txxpv\" (UniqueName: \"kubernetes.io/projected/20e7b7c5-1480-498e-ab4d-2e335da52dd3-kube-api-access-txxpv\") pod \"machine-approver-56656f9798-6jzdp\" (UID: \"20e7b7c5-1480-498e-ab4d-2e335da52dd3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jzdp" Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.449466 4998 configmap.go:193] Couldn't get configMap openshift-service-ca/signing-cabundle: failed to sync configmap cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.449583 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/75498a9e-2bcc-4e76-a528-16df0f6d8ae7-signing-cabundle podName:75498a9e-2bcc-4e76-a528-16df0f6d8ae7 nodeName:}" failed. No retries permitted until 2025-12-03 16:05:09.949552113 +0000 UTC m=+88.561252346 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-cabundle" (UniqueName: "kubernetes.io/configmap/75498a9e-2bcc-4e76-a528-16df0f6d8ae7-signing-cabundle") pod "service-ca-9c57cc56f-v8c75" (UID: "75498a9e-2bcc-4e76-a528-16df0f6d8ae7") : failed to sync configmap cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.449616 4998 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.449655 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7274d4ec-5476-46c7-ab6a-f4ed0f4b294e-package-server-manager-serving-cert podName:7274d4ec-5476-46c7-ab6a-f4ed0f4b294e nodeName:}" failed. No retries permitted until 2025-12-03 16:05:09.949644315 +0000 UTC m=+88.561344548 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/7274d4ec-5476-46c7-ab6a-f4ed0f4b294e-package-server-manager-serving-cert") pod "package-server-manager-789f6589d5-jmdsz" (UID: "7274d4ec-5476-46c7-ab6a-f4ed0f4b294e") : failed to sync secret cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.449692 4998 configmap.go:193] Couldn't get configMap openshift-etcd-operator/etcd-operator-config: failed to sync configmap cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.449727 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-config podName:77eb0bc5-1a4a-48ae-98cc-f26a450af04a nodeName:}" failed. No retries permitted until 2025-12-03 16:05:09.949716317 +0000 UTC m=+88.561416560 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-config") pod "etcd-operator-b45778765-vns7l" (UID: "77eb0bc5-1a4a-48ae-98cc-f26a450af04a") : failed to sync configmap cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.449770 4998 secret.go:188] Couldn't get secret openshift-etcd-operator/etcd-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.449804 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-serving-cert podName:77eb0bc5-1a4a-48ae-98cc-f26a450af04a nodeName:}" failed. No retries permitted until 2025-12-03 16:05:09.949793568 +0000 UTC m=+88.561493801 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-serving-cert") pod "etcd-operator-b45778765-vns7l" (UID: "77eb0bc5-1a4a-48ae-98cc-f26a450af04a") : failed to sync secret cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.449843 4998 secret.go:188] Couldn't get secret openshift-dns/dns-default-metrics-tls: failed to sync secret cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.449835 4998 configmap.go:193] Couldn't get configMap openshift-dns/dns-default: failed to sync configmap cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.449877 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a741e843-b85d-4250-844a-b03534dc47d2-metrics-tls podName:a741e843-b85d-4250-844a-b03534dc47d2 nodeName:}" failed. No retries permitted until 2025-12-03 16:05:09.94986855 +0000 UTC m=+88.561568783 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/a741e843-b85d-4250-844a-b03534dc47d2-metrics-tls") pod "dns-default-wsmkz" (UID: "a741e843-b85d-4250-844a-b03534dc47d2") : failed to sync secret cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.449905 4998 secret.go:188] Couldn't get secret openshift-service-ca/signing-key: failed to sync secret cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.449937 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/75498a9e-2bcc-4e76-a528-16df0f6d8ae7-signing-key podName:75498a9e-2bcc-4e76-a528-16df0f6d8ae7 nodeName:}" failed. No retries permitted until 2025-12-03 16:05:09.949928171 +0000 UTC m=+88.561628404 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-key" (UniqueName: "kubernetes.io/secret/75498a9e-2bcc-4e76-a528-16df0f6d8ae7-signing-key") pod "service-ca-9c57cc56f-v8c75" (UID: "75498a9e-2bcc-4e76-a528-16df0f6d8ae7") : failed to sync secret cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.449967 4998 secret.go:188] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: failed to sync secret cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.450000 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c715a96-664c-4b26-8d5a-523a0fb4a17c-marketplace-operator-metrics podName:8c715a96-664c-4b26-8d5a-523a0fb4a17c nodeName:}" failed. No retries permitted until 2025-12-03 16:05:09.949991973 +0000 UTC m=+88.561692206 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/8c715a96-664c-4b26-8d5a-523a0fb4a17c-marketplace-operator-metrics") pod "marketplace-operator-79b997595-fnxt7" (UID: "8c715a96-664c-4b26-8d5a-523a0fb4a17c") : failed to sync secret cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.450052 4998 configmap.go:193] Couldn't get configMap openshift-etcd-operator/etcd-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.450144 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-etcd-ca podName:77eb0bc5-1a4a-48ae-98cc-f26a450af04a nodeName:}" failed. No retries permitted until 2025-12-03 16:05:09.950106295 +0000 UTC m=+88.561806578 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-ca" (UniqueName: "kubernetes.io/configmap/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-etcd-ca") pod "etcd-operator-b45778765-vns7l" (UID: "77eb0bc5-1a4a-48ae-98cc-f26a450af04a") : failed to sync configmap cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.450150 4998 secret.go:188] Couldn't get secret openshift-etcd-operator/etcd-client: failed to sync secret cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.450221 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-etcd-client podName:77eb0bc5-1a4a-48ae-98cc-f26a450af04a nodeName:}" failed. No retries permitted until 2025-12-03 16:05:09.950200317 +0000 UTC m=+88.561900580 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-etcd-client") pod "etcd-operator-b45778765-vns7l" (UID: "77eb0bc5-1a4a-48ae-98cc-f26a450af04a") : failed to sync secret cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.450285 4998 configmap.go:193] Couldn't get configMap openshift-marketplace/marketplace-trusted-ca: failed to sync configmap cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.450330 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8c715a96-664c-4b26-8d5a-523a0fb4a17c-marketplace-trusted-ca podName:8c715a96-664c-4b26-8d5a-523a0fb4a17c nodeName:}" failed. No retries permitted until 2025-12-03 16:05:09.95031431 +0000 UTC m=+88.562014573 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-trusted-ca" (UniqueName: "kubernetes.io/configmap/8c715a96-664c-4b26-8d5a-523a0fb4a17c-marketplace-trusted-ca") pod "marketplace-operator-79b997595-fnxt7" (UID: "8c715a96-664c-4b26-8d5a-523a0fb4a17c") : failed to sync configmap cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.450361 4998 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.450399 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bff7045b-cac0-438c-b6a6-769147c3f108-apiservice-cert podName:bff7045b-cac0-438c-b6a6-769147c3f108 nodeName:}" failed. No retries permitted until 2025-12-03 16:05:09.950387481 +0000 UTC m=+88.562087744 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/bff7045b-cac0-438c-b6a6-769147c3f108-apiservice-cert") pod "packageserver-d55dfcdfc-zh6tw" (UID: "bff7045b-cac0-438c-b6a6-769147c3f108") : failed to sync secret cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.450421 4998 secret.go:188] Couldn't get secret openshift-kube-storage-version-migrator-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.450458 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/02311a5a-a912-4cce-b076-d752266f0c95-serving-cert podName:02311a5a-a912-4cce-b076-d752266f0c95 nodeName:}" failed. No retries permitted until 2025-12-03 16:05:09.950447763 +0000 UTC m=+88.562148036 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/02311a5a-a912-4cce-b076-d752266f0c95-serving-cert") pod "kube-storage-version-migrator-operator-b67b599dd-8l9b5" (UID: "02311a5a-a912-4cce-b076-d752266f0c95") : failed to sync secret cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.450508 4998 configmap.go:193] Couldn't get configMap openshift-kube-storage-version-migrator-operator/config: failed to sync configmap cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.450531 4998 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.450545 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/02311a5a-a912-4cce-b076-d752266f0c95-config podName:02311a5a-a912-4cce-b076-d752266f0c95 nodeName:}" failed. No retries permitted until 2025-12-03 16:05:09.950533485 +0000 UTC m=+88.562233748 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/02311a5a-a912-4cce-b076-d752266f0c95-config") pod "kube-storage-version-migrator-operator-b67b599dd-8l9b5" (UID: "02311a5a-a912-4cce-b076-d752266f0c95") : failed to sync configmap cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.450574 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bff7045b-cac0-438c-b6a6-769147c3f108-webhook-cert podName:bff7045b-cac0-438c-b6a6-769147c3f108 nodeName:}" failed. No retries permitted until 2025-12-03 16:05:09.950564085 +0000 UTC m=+88.562264318 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/bff7045b-cac0-438c-b6a6-769147c3f108-webhook-cert") pod "packageserver-d55dfcdfc-zh6tw" (UID: "bff7045b-cac0-438c-b6a6-769147c3f108") : failed to sync secret cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.450605 4998 secret.go:188] Couldn't get secret openshift-machine-api/control-plane-machine-set-operator-tls: failed to sync secret cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.450636 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12d866d2-1002-4db4-9832-59b5fea5062e-control-plane-machine-set-operator-tls podName:12d866d2-1002-4db4-9832-59b5fea5062e nodeName:}" failed. No retries permitted until 2025-12-03 16:05:09.950627637 +0000 UTC m=+88.562327870 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "control-plane-machine-set-operator-tls" (UniqueName: "kubernetes.io/secret/12d866d2-1002-4db4-9832-59b5fea5062e-control-plane-machine-set-operator-tls") pod "control-plane-machine-set-operator-78cbb6b69f-sn2r4" (UID: "12d866d2-1002-4db4-9832-59b5fea5062e") : failed to sync secret cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.450908 4998 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.450963 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a6e6bbd8-c854-4708-b9b1-40884d1be6c7-config-volume podName:a6e6bbd8-c854-4708-b9b1-40884d1be6c7 nodeName:}" failed. No retries permitted until 2025-12-03 16:05:09.950947844 +0000 UTC m=+88.562648077 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/a6e6bbd8-c854-4708-b9b1-40884d1be6c7-config-volume") pod "collect-profiles-29412960-6vr8n" (UID: "a6e6bbd8-c854-4708-b9b1-40884d1be6c7") : failed to sync configmap cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.451007 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a741e843-b85d-4250-844a-b03534dc47d2-config-volume podName:a741e843-b85d-4250-844a-b03534dc47d2 nodeName:}" failed. No retries permitted until 2025-12-03 16:05:09.950993215 +0000 UTC m=+88.562693448 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/a741e843-b85d-4250-844a-b03534dc47d2-config-volume") pod "dns-default-wsmkz" (UID: "a741e843-b85d-4250-844a-b03534dc47d2") : failed to sync configmap cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.451133 4998 secret.go:188] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.451181 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6ce38b01-0835-4a45-a262-f584fe51bd62-cert podName:6ce38b01-0835-4a45-a262-f584fe51bd62 nodeName:}" failed. No retries permitted until 2025-12-03 16:05:09.951168489 +0000 UTC m=+88.562868722 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6ce38b01-0835-4a45-a262-f584fe51bd62-cert") pod "ingress-canary-b9f5r" (UID: "6ce38b01-0835-4a45-a262-f584fe51bd62") : failed to sync secret cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.451222 4998 configmap.go:193] Couldn't get configMap openshift-etcd-operator/etcd-service-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: E1203 16:05:09.451258 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-etcd-service-ca podName:77eb0bc5-1a4a-48ae-98cc-f26a450af04a nodeName:}" failed. No retries permitted until 2025-12-03 16:05:09.95124699 +0000 UTC m=+88.562947393 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-service-ca" (UniqueName: "kubernetes.io/configmap/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-etcd-service-ca") pod "etcd-operator-b45778765-vns7l" (UID: "77eb0bc5-1a4a-48ae-98cc-f26a450af04a") : failed to sync configmap cache: timed out waiting for the condition Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.455564 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4dj2\" (UniqueName: \"kubernetes.io/projected/94922a63-40d9-4944-9d95-f860967ea5b3-kube-api-access-d4dj2\") pod \"cluster-samples-operator-665b6dd947-9cdmt\" (UID: \"94922a63-40d9-4944-9d95-f860967ea5b3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9cdmt" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.460958 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.475375 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.497889 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.527578 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.535990 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.556115 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.575919 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.596236 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.615807 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.618344 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6"] Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.633467 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.634774 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.651046 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-hwftp"] Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.655551 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 16:05:09 crc kubenswrapper[4998]: W1203 16:05:09.660728 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda556e01d_7f37_46db_ac77_6a681cad9e7e.slice/crio-bb7d30258a1599bf52ae098f28d9b3e6b04d202c9c6aef6d1e6c576c9e3738a7 WatchSource:0}: Error finding container bb7d30258a1599bf52ae098f28d9b3e6b04d202c9c6aef6d1e6c576c9e3738a7: Status 404 returned error can't find the container with id bb7d30258a1599bf52ae098f28d9b3e6b04d202c9c6aef6d1e6c576c9e3738a7 Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.663200 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jzdp" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.676182 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.676730 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:05:09 crc kubenswrapper[4998]: W1203 16:05:09.683313 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20e7b7c5_1480_498e_ab4d_2e335da52dd3.slice/crio-548ffeae60a128edd255e1c76d05aa16f8dfbbc9f3ddc902777f5b085a5a1c43 WatchSource:0}: Error finding container 548ffeae60a128edd255e1c76d05aa16f8dfbbc9f3ddc902777f5b085a5a1c43: Status 404 returned error can't find the container with id 548ffeae60a128edd255e1c76d05aa16f8dfbbc9f3ddc902777f5b085a5a1c43 Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.695593 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.703586 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9cdmt" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.716508 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.735761 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.755392 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.775326 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.788947 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2kprn"] Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.795100 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.816338 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.835383 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.855275 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.875868 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.889470 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9cdmt"] Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.895226 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.915379 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.935829 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.955166 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.975295 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.979890 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02311a5a-a912-4cce-b076-d752266f0c95-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8l9b5\" (UID: \"02311a5a-a912-4cce-b076-d752266f0c95\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l9b5" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.979917 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bff7045b-cac0-438c-b6a6-769147c3f108-apiservice-cert\") pod \"packageserver-d55dfcdfc-zh6tw\" (UID: \"bff7045b-cac0-438c-b6a6-769147c3f108\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zh6tw" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.979948 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02311a5a-a912-4cce-b076-d752266f0c95-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8l9b5\" (UID: \"02311a5a-a912-4cce-b076-d752266f0c95\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l9b5" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.979965 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bff7045b-cac0-438c-b6a6-769147c3f108-webhook-cert\") pod \"packageserver-d55dfcdfc-zh6tw\" (UID: \"bff7045b-cac0-438c-b6a6-769147c3f108\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zh6tw" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.980046 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-etcd-service-ca\") pod \"etcd-operator-b45778765-vns7l\" (UID: \"77eb0bc5-1a4a-48ae-98cc-f26a450af04a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.980091 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/75498a9e-2bcc-4e76-a528-16df0f6d8ae7-signing-cabundle\") pod \"service-ca-9c57cc56f-v8c75\" (UID: \"75498a9e-2bcc-4e76-a528-16df0f6d8ae7\") " pod="openshift-service-ca/service-ca-9c57cc56f-v8c75" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.980117 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6ce38b01-0835-4a45-a262-f584fe51bd62-cert\") pod \"ingress-canary-b9f5r\" (UID: \"6ce38b01-0835-4a45-a262-f584fe51bd62\") " pod="openshift-ingress-canary/ingress-canary-b9f5r" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.980136 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7274d4ec-5476-46c7-ab6a-f4ed0f4b294e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jmdsz\" (UID: \"7274d4ec-5476-46c7-ab6a-f4ed0f4b294e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jmdsz" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.980155 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a741e843-b85d-4250-844a-b03534dc47d2-metrics-tls\") pod \"dns-default-wsmkz\" (UID: \"a741e843-b85d-4250-844a-b03534dc47d2\") " pod="openshift-dns/dns-default-wsmkz" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.980201 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/75498a9e-2bcc-4e76-a528-16df0f6d8ae7-signing-key\") pod \"service-ca-9c57cc56f-v8c75\" (UID: \"75498a9e-2bcc-4e76-a528-16df0f6d8ae7\") " pod="openshift-service-ca/service-ca-9c57cc56f-v8c75" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.980216 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-etcd-client\") pod \"etcd-operator-b45778765-vns7l\" (UID: \"77eb0bc5-1a4a-48ae-98cc-f26a450af04a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.980234 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-serving-cert\") pod \"etcd-operator-b45778765-vns7l\" (UID: \"77eb0bc5-1a4a-48ae-98cc-f26a450af04a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.980261 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-config\") pod \"etcd-operator-b45778765-vns7l\" (UID: \"77eb0bc5-1a4a-48ae-98cc-f26a450af04a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.980311 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a741e843-b85d-4250-844a-b03534dc47d2-config-volume\") pod \"dns-default-wsmkz\" (UID: \"a741e843-b85d-4250-844a-b03534dc47d2\") " pod="openshift-dns/dns-default-wsmkz" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.980344 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-etcd-ca\") pod \"etcd-operator-b45778765-vns7l\" (UID: \"77eb0bc5-1a4a-48ae-98cc-f26a450af04a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.980384 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8c715a96-664c-4b26-8d5a-523a0fb4a17c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-fnxt7\" (UID: \"8c715a96-664c-4b26-8d5a-523a0fb4a17c\") " pod="openshift-marketplace/marketplace-operator-79b997595-fnxt7" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.980401 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a6e6bbd8-c854-4708-b9b1-40884d1be6c7-config-volume\") pod \"collect-profiles-29412960-6vr8n\" (UID: \"a6e6bbd8-c854-4708-b9b1-40884d1be6c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-6vr8n" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.980449 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c715a96-664c-4b26-8d5a-523a0fb4a17c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-fnxt7\" (UID: \"8c715a96-664c-4b26-8d5a-523a0fb4a17c\") " pod="openshift-marketplace/marketplace-operator-79b997595-fnxt7" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.980465 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/12d866d2-1002-4db4-9832-59b5fea5062e-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-sn2r4\" (UID: \"12d866d2-1002-4db4-9832-59b5fea5062e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sn2r4" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.981279 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02311a5a-a912-4cce-b076-d752266f0c95-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8l9b5\" (UID: \"02311a5a-a912-4cce-b076-d752266f0c95\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l9b5" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.981309 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-etcd-ca\") pod \"etcd-operator-b45778765-vns7l\" (UID: \"77eb0bc5-1a4a-48ae-98cc-f26a450af04a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.982316 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/75498a9e-2bcc-4e76-a528-16df0f6d8ae7-signing-cabundle\") pod \"service-ca-9c57cc56f-v8c75\" (UID: \"75498a9e-2bcc-4e76-a528-16df0f6d8ae7\") " pod="openshift-service-ca/service-ca-9c57cc56f-v8c75" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.982635 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a6e6bbd8-c854-4708-b9b1-40884d1be6c7-config-volume\") pod \"collect-profiles-29412960-6vr8n\" (UID: \"a6e6bbd8-c854-4708-b9b1-40884d1be6c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-6vr8n" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.983099 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c715a96-664c-4b26-8d5a-523a0fb4a17c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-fnxt7\" (UID: \"8c715a96-664c-4b26-8d5a-523a0fb4a17c\") " pod="openshift-marketplace/marketplace-operator-79b997595-fnxt7" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.984417 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/75498a9e-2bcc-4e76-a528-16df0f6d8ae7-signing-key\") pod \"service-ca-9c57cc56f-v8c75\" (UID: \"75498a9e-2bcc-4e76-a528-16df0f6d8ae7\") " pod="openshift-service-ca/service-ca-9c57cc56f-v8c75" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.984421 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-etcd-client\") pod \"etcd-operator-b45778765-vns7l\" (UID: \"77eb0bc5-1a4a-48ae-98cc-f26a450af04a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.984730 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8c715a96-664c-4b26-8d5a-523a0fb4a17c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-fnxt7\" (UID: \"8c715a96-664c-4b26-8d5a-523a0fb4a17c\") " pod="openshift-marketplace/marketplace-operator-79b997595-fnxt7" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.985204 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02311a5a-a912-4cce-b076-d752266f0c95-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8l9b5\" (UID: \"02311a5a-a912-4cce-b076-d752266f0c95\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l9b5" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.985639 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bff7045b-cac0-438c-b6a6-769147c3f108-apiservice-cert\") pod \"packageserver-d55dfcdfc-zh6tw\" (UID: \"bff7045b-cac0-438c-b6a6-769147c3f108\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zh6tw" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.986038 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-serving-cert\") pod \"etcd-operator-b45778765-vns7l\" (UID: \"77eb0bc5-1a4a-48ae-98cc-f26a450af04a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.986559 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bff7045b-cac0-438c-b6a6-769147c3f108-webhook-cert\") pod \"packageserver-d55dfcdfc-zh6tw\" (UID: \"bff7045b-cac0-438c-b6a6-769147c3f108\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zh6tw" Dec 03 16:05:09 crc kubenswrapper[4998]: I1203 16:05:09.997145 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.002159 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-etcd-service-ca\") pod \"etcd-operator-b45778765-vns7l\" (UID: \"77eb0bc5-1a4a-48ae-98cc-f26a450af04a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.015901 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.022535 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-config\") pod \"etcd-operator-b45778765-vns7l\" (UID: \"77eb0bc5-1a4a-48ae-98cc-f26a450af04a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.035910 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.055928 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.065390 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/12d866d2-1002-4db4-9832-59b5fea5062e-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-sn2r4\" (UID: \"12d866d2-1002-4db4-9832-59b5fea5062e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sn2r4" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.075304 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.086208 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7274d4ec-5476-46c7-ab6a-f4ed0f4b294e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jmdsz\" (UID: \"7274d4ec-5476-46c7-ab6a-f4ed0f4b294e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jmdsz" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.096189 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.116121 4998 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.136343 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.155674 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.167573 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6ce38b01-0835-4a45-a262-f584fe51bd62-cert\") pod \"ingress-canary-b9f5r\" (UID: \"6ce38b01-0835-4a45-a262-f584fe51bd62\") " pod="openshift-ingress-canary/ingress-canary-b9f5r" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.176106 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.195791 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.213663 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hwftp" event={"ID":"a556e01d-7f37-46db-ac77-6a681cad9e7e","Type":"ContainerStarted","Data":"bb7d30258a1599bf52ae098f28d9b3e6b04d202c9c6aef6d1e6c576c9e3738a7"} Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.215392 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.216127 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" event={"ID":"c2ff7dad-a4da-4243-ab1e-21b47230504d","Type":"ContainerStarted","Data":"1debfca7c9ac2dd15259c876cdeaf36a666d9b1e9969b08ad0388c33ae038c76"} Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.218029 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jzdp" event={"ID":"20e7b7c5-1480-498e-ab4d-2e335da52dd3","Type":"ContainerStarted","Data":"548ffeae60a128edd255e1c76d05aa16f8dfbbc9f3ddc902777f5b085a5a1c43"} Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.219715 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" event={"ID":"b46224fb-c151-4ba9-87c7-85e0ed6e11bc","Type":"ContainerStarted","Data":"6a7f45d27042c4bd1b8fa5370666963a8af0b2023f0d0353619c6a5b27f315b6"} Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.235198 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.243277 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a741e843-b85d-4250-844a-b03534dc47d2-config-volume\") pod \"dns-default-wsmkz\" (UID: \"a741e843-b85d-4250-844a-b03534dc47d2\") " pod="openshift-dns/dns-default-wsmkz" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.256090 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.273850 4998 request.go:700] Waited for 1.964222478s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns/secrets?fieldSelector=metadata.name%3Ddns-default-metrics-tls&limit=500&resourceVersion=0 Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.276068 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.288311 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a741e843-b85d-4250-844a-b03534dc47d2-metrics-tls\") pod \"dns-default-wsmkz\" (UID: \"a741e843-b85d-4250-844a-b03534dc47d2\") " pod="openshift-dns/dns-default-wsmkz" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.296295 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.316101 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.335557 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.402739 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rhgj\" (UniqueName: \"kubernetes.io/projected/c71a13b3-a727-4b88-9be8-954a62dc7c86-kube-api-access-4rhgj\") pod \"ingress-operator-5b745b69d9-8ft9r\" (UID: \"c71a13b3-a727-4b88-9be8-954a62dc7c86\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8ft9r" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.420126 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm8zs\" (UniqueName: \"kubernetes.io/projected/7a975a4d-12df-4c73-b644-ff281b3c3390-kube-api-access-tm8zs\") pod \"cluster-image-registry-operator-dc59b4c8b-8lmkd\" (UID: \"7a975a4d-12df-4c73-b644-ff281b3c3390\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8lmkd" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.443434 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2r7z\" (UniqueName: \"kubernetes.io/projected/56ad3380-4861-49d1-8758-7b1e27f74560-kube-api-access-n2r7z\") pod \"oauth-openshift-558db77b4-b5hpf\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.452131 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e569f94d-9231-4277-af77-f97c23fcbabe-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-wk6qg\" (UID: \"e569f94d-9231-4277-af77-f97c23fcbabe\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wk6qg" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.489835 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c71a13b3-a727-4b88-9be8-954a62dc7c86-bound-sa-token\") pod \"ingress-operator-5b745b69d9-8ft9r\" (UID: \"c71a13b3-a727-4b88-9be8-954a62dc7c86\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8ft9r" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.502925 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.509437 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7a975a4d-12df-4c73-b644-ff281b3c3390-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8lmkd\" (UID: \"7a975a4d-12df-4c73-b644-ff281b3c3390\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8lmkd" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.542209 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjd2p\" (UniqueName: \"kubernetes.io/projected/03021110-7732-4dff-a5e3-f481e8b7c0d6-kube-api-access-pjd2p\") pod \"console-f9d7485db-kk8mb\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.554380 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68597ef9-34cf-45c4-86c7-153f9f4d9363-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5wc9c\" (UID: \"68597ef9-34cf-45c4-86c7-153f9f4d9363\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5wc9c" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.563123 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zk89t\" (UniqueName: \"kubernetes.io/projected/892e64fc-db6b-4410-9b4c-3ff78cc82074-kube-api-access-zk89t\") pod \"machine-api-operator-5694c8668f-ggw6k\" (UID: \"892e64fc-db6b-4410-9b4c-3ff78cc82074\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ggw6k" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.579201 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84e61bdd-8de2-4cd1-93a2-b43d802fbea3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-ctpbx\" (UID: \"84e61bdd-8de2-4cd1-93a2-b43d802fbea3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctpbx" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.589992 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8ft9r" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.610178 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7kj8\" (UniqueName: \"kubernetes.io/projected/2e58b61f-82bf-42c2-a664-cd3fcbd0fc03-kube-api-access-l7kj8\") pod \"downloads-7954f5f757-ppw5v\" (UID: \"2e58b61f-82bf-42c2-a664-cd3fcbd0fc03\") " pod="openshift-console/downloads-7954f5f757-ppw5v" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.616292 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptcqb\" (UniqueName: \"kubernetes.io/projected/ac8c6176-0450-4050-9094-003067cd40b1-kube-api-access-ptcqb\") pod \"openshift-config-operator-7777fb866f-kg4lw\" (UID: \"ac8c6176-0450-4050-9094-003067cd40b1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kg4lw" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.636044 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7r6j\" (UniqueName: \"kubernetes.io/projected/e1ff9891-7017-41a2-b7c0-183caa783671-kube-api-access-g7r6j\") pod \"console-operator-58897d9998-qvmlb\" (UID: \"e1ff9891-7017-41a2-b7c0-183caa783671\") " pod="openshift-console-operator/console-operator-58897d9998-qvmlb" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.653076 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqpn4\" (UniqueName: \"kubernetes.io/projected/09d80f4a-297b-4000-8fcd-c586b5267339-kube-api-access-gqpn4\") pod \"openshift-apiserver-operator-796bbdcf4f-nsq9q\" (UID: \"09d80f4a-297b-4000-8fcd-c586b5267339\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nsq9q" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.676875 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkxbv\" (UniqueName: \"kubernetes.io/projected/0a1b6dae-ff3d-4bfb-a811-98866c9c2b59-kube-api-access-hkxbv\") pod \"openshift-controller-manager-operator-756b6f6bc6-54vrz\" (UID: \"0a1b6dae-ff3d-4bfb-a811-98866c9c2b59\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-54vrz" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.679352 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nsq9q" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.686904 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wk6qg" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.694249 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-qvmlb" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.695829 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzn2p\" (UniqueName: \"kubernetes.io/projected/6024dcd8-96d6-4c42-9670-42e3608cd7ee-kube-api-access-tzn2p\") pod \"route-controller-manager-6576b87f9c-tvqq9\" (UID: \"6024dcd8-96d6-4c42-9670-42e3608cd7ee\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.701438 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kg4lw" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.709058 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-ggw6k" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.710706 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-b5hpf"] Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.711469 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wst5d\" (UniqueName: \"kubernetes.io/projected/22c37be4-5e07-4638-aa01-04ee06c463d7-kube-api-access-wst5d\") pod \"multus-admission-controller-857f4d67dd-r2gbh\" (UID: \"22c37be4-5e07-4638-aa01-04ee06c463d7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-r2gbh" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.719158 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctpbx" Dec 03 16:05:10 crc kubenswrapper[4998]: W1203 16:05:10.722407 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56ad3380_4861_49d1_8758_7b1e27f74560.slice/crio-f6b1e607b8be12fd2922ddb6c59e52604deb2fdad5903c35d87e7586f4f82867 WatchSource:0}: Error finding container f6b1e607b8be12fd2922ddb6c59e52604deb2fdad5903c35d87e7586f4f82867: Status 404 returned error can't find the container with id f6b1e607b8be12fd2922ddb6c59e52604deb2fdad5903c35d87e7586f4f82867 Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.724667 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.737092 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-ppw5v" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.737515 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8hgn\" (UniqueName: \"kubernetes.io/projected/57c54ccc-a846-4c00-8318-5d5331c70e70-kube-api-access-q8hgn\") pod \"dns-operator-744455d44c-4rq7l\" (UID: \"57c54ccc-a846-4c00-8318-5d5331c70e70\") " pod="openshift-dns-operator/dns-operator-744455d44c-4rq7l" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.744379 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-54vrz" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.749467 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxtbv\" (UniqueName: \"kubernetes.io/projected/235934fb-ec04-4784-b30c-5c80dfa325ad-kube-api-access-bxtbv\") pod \"authentication-operator-69f744f599-5c2lp\" (UID: \"235934fb-ec04-4784-b30c-5c80dfa325ad\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5c2lp" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.776677 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w5qv\" (UniqueName: \"kubernetes.io/projected/8c88814f-7a99-4878-9517-55cbecd72ff0-kube-api-access-9w5qv\") pod \"machine-config-controller-84d6567774-jjwjb\" (UID: \"8c88814f-7a99-4878-9517-55cbecd72ff0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jjwjb" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.782043 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5wc9c" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.787909 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8lmkd" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.793925 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbj97\" (UniqueName: \"kubernetes.io/projected/a6e6bbd8-c854-4708-b9b1-40884d1be6c7-kube-api-access-lbj97\") pod \"collect-profiles-29412960-6vr8n\" (UID: \"a6e6bbd8-c854-4708-b9b1-40884d1be6c7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-6vr8n" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.794960 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-r2gbh" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.796706 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-8ft9r"] Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.810180 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2dvn\" (UniqueName: \"kubernetes.io/projected/6ce38b01-0835-4a45-a262-f584fe51bd62-kube-api-access-c2dvn\") pod \"ingress-canary-b9f5r\" (UID: \"6ce38b01-0835-4a45-a262-f584fe51bd62\") " pod="openshift-ingress-canary/ingress-canary-b9f5r" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.827574 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a29a797b-0ffe-4e7c-b4c4-d3869fb694fe-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rrj4k\" (UID: \"a29a797b-0ffe-4e7c-b4c4-d3869fb694fe\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rrj4k" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.847822 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wk6qg"] Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.850179 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xjhl\" (UniqueName: \"kubernetes.io/projected/77eb0bc5-1a4a-48ae-98cc-f26a450af04a-kube-api-access-9xjhl\") pod \"etcd-operator-b45778765-vns7l\" (UID: \"77eb0bc5-1a4a-48ae-98cc-f26a450af04a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.852062 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jjwjb" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.871699 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dvph\" (UniqueName: \"kubernetes.io/projected/12d866d2-1002-4db4-9832-59b5fea5062e-kube-api-access-7dvph\") pod \"control-plane-machine-set-operator-78cbb6b69f-sn2r4\" (UID: \"12d866d2-1002-4db4-9832-59b5fea5062e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sn2r4" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.882250 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nsq9q"] Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.889317 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44wgw\" (UniqueName: \"kubernetes.io/projected/a741e843-b85d-4250-844a-b03534dc47d2-kube-api-access-44wgw\") pod \"dns-default-wsmkz\" (UID: \"a741e843-b85d-4250-844a-b03534dc47d2\") " pod="openshift-dns/dns-default-wsmkz" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.906338 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-6vr8n" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.910659 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs7mz\" (UniqueName: \"kubernetes.io/projected/f089be95-2900-436d-b1f4-627095e2d563-kube-api-access-bs7mz\") pod \"machine-config-operator-74547568cd-dc55h\" (UID: \"f089be95-2900-436d-b1f4-627095e2d563\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dc55h" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.914671 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" Dec 03 16:05:10 crc kubenswrapper[4998]: W1203 16:05:10.916893 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84e61bdd_8de2_4cd1_93a2_b43d802fbea3.slice/crio-808664cea365de4f748218334dacb9a55daefc318b8602fc9ee740583ad4b0cd WatchSource:0}: Error finding container 808664cea365de4f748218334dacb9a55daefc318b8602fc9ee740583ad4b0cd: Status 404 returned error can't find the container with id 808664cea365de4f748218334dacb9a55daefc318b8602fc9ee740583ad4b0cd Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.924541 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sn2r4" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.936136 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7lbl\" (UniqueName: \"kubernetes.io/projected/ef0bf68e-f7d5-43a2-890c-10ffba3e6238-kube-api-access-q7lbl\") pod \"migrator-59844c95c7-bhj2p\" (UID: \"ef0bf68e-f7d5-43a2-890c-10ffba3e6238\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bhj2p" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.940044 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-4rq7l" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.949180 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frnjz\" (UniqueName: \"kubernetes.io/projected/9a6a4ae3-f219-4a1b-bfbe-f9f829945a8a-kube-api-access-frnjz\") pod \"catalog-operator-68c6474976-xd7m2\" (UID: \"9a6a4ae3-f219-4a1b-bfbe-f9f829945a8a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xd7m2" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.949435 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.969888 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxzlg\" (UniqueName: \"kubernetes.io/projected/31bfedfc-39de-4bf0-b049-e41c577ac848-kube-api-access-nxzlg\") pod \"olm-operator-6b444d44fb-wfm8s\" (UID: \"31bfedfc-39de-4bf0-b049-e41c577ac848\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wfm8s" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.983407 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-b9f5r" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.990414 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-555gk\" (UniqueName: \"kubernetes.io/projected/cbae1035-ec59-4f24-884d-809c7e7bb685-kube-api-access-555gk\") pod \"router-default-5444994796-hnvvp\" (UID: \"cbae1035-ec59-4f24-884d-809c7e7bb685\") " pod="openshift-ingress/router-default-5444994796-hnvvp" Dec 03 16:05:10 crc kubenswrapper[4998]: I1203 16:05:10.991866 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-wsmkz" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.015282 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zvj2\" (UniqueName: \"kubernetes.io/projected/bff7045b-cac0-438c-b6a6-769147c3f108-kube-api-access-4zvj2\") pod \"packageserver-d55dfcdfc-zh6tw\" (UID: \"bff7045b-cac0-438c-b6a6-769147c3f108\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zh6tw" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.022550 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-ggw6k"] Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.030875 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-5c2lp" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.040911 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gt8xm\" (UniqueName: \"kubernetes.io/projected/8c715a96-664c-4b26-8d5a-523a0fb4a17c-kube-api-access-gt8xm\") pod \"marketplace-operator-79b997595-fnxt7\" (UID: \"8c715a96-664c-4b26-8d5a-523a0fb4a17c\") " pod="openshift-marketplace/marketplace-operator-79b997595-fnxt7" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.058681 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qmnv\" (UniqueName: \"kubernetes.io/projected/e7a61bf3-1760-4910-9561-000d68ca9708-kube-api-access-6qmnv\") pod \"csi-hostpathplugin-5mthf\" (UID: \"e7a61bf3-1760-4910-9561-000d68ca9708\") " pod="hostpath-provisioner/csi-hostpathplugin-5mthf" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.070101 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m985g\" (UniqueName: \"kubernetes.io/projected/817c8b8a-cab9-474a-ac82-30b55677c2f4-kube-api-access-m985g\") pod \"service-ca-operator-777779d784-k92zl\" (UID: \"817c8b8a-cab9-474a-ac82-30b55677c2f4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k92zl" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.080004 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-kg4lw"] Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.086248 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qvmlb"] Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.104300 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srczp\" (UniqueName: \"kubernetes.io/projected/7274d4ec-5476-46c7-ab6a-f4ed0f4b294e-kube-api-access-srczp\") pod \"package-server-manager-789f6589d5-jmdsz\" (UID: \"7274d4ec-5476-46c7-ab6a-f4ed0f4b294e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jmdsz" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.110644 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rrj4k" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.117743 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nd5s5\" (UniqueName: \"kubernetes.io/projected/75498a9e-2bcc-4e76-a528-16df0f6d8ae7-kube-api-access-nd5s5\") pod \"service-ca-9c57cc56f-v8c75\" (UID: \"75498a9e-2bcc-4e76-a528-16df0f6d8ae7\") " pod="openshift-service-ca/service-ca-9c57cc56f-v8c75" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.125075 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xd7m2" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.130687 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dc55h" Dec 03 16:05:11 crc kubenswrapper[4998]: W1203 16:05:11.131335 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac8c6176_0450_4050_9094_003067cd40b1.slice/crio-c10cc4f9c0a7b079b7603306e605afdce7d29e1b9f63c886deb04ac661bcba66 WatchSource:0}: Error finding container c10cc4f9c0a7b079b7603306e605afdce7d29e1b9f63c886deb04ac661bcba66: Status 404 returned error can't find the container with id c10cc4f9c0a7b079b7603306e605afdce7d29e1b9f63c886deb04ac661bcba66 Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.135672 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.136183 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9lqv\" (UniqueName: \"kubernetes.io/projected/02311a5a-a912-4cce-b076-d752266f0c95-kube-api-access-v9lqv\") pod \"kube-storage-version-migrator-operator-b67b599dd-8l9b5\" (UID: \"02311a5a-a912-4cce-b076-d752266f0c95\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l9b5" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.137127 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-hnvvp" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.144305 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wfm8s" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.155386 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.165472 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-k92zl" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.168589 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-fnxt7" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.176500 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.176914 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bhj2p" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.184458 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l9b5" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.192228 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zh6tw" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.195072 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.199157 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-v8c75" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.234170 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jmdsz" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.235112 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.241183 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9cdmt" event={"ID":"94922a63-40d9-4944-9d95-f860967ea5b3","Type":"ContainerStarted","Data":"e0f6ab81b25f8e75ffca2719bb565fa2ae12404d42bd7b2538390b2e647d9147"} Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.243692 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8lmkd"] Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.257152 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.268503 4998 generic.go:334] "Generic (PLEG): container finished" podID="b46224fb-c151-4ba9-87c7-85e0ed6e11bc" containerID="dca8c36da4a6c3eba79cc4c13507f38bd132258d19c2754b77629161d9e57b9b" exitCode=0 Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.268694 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" event={"ID":"b46224fb-c151-4ba9-87c7-85e0ed6e11bc","Type":"ContainerDied","Data":"dca8c36da4a6c3eba79cc4c13507f38bd132258d19c2754b77629161d9e57b9b"} Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.276675 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-5mthf" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.288966 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kg4lw" event={"ID":"ac8c6176-0450-4050-9094-003067cd40b1","Type":"ContainerStarted","Data":"c10cc4f9c0a7b079b7603306e605afdce7d29e1b9f63c886deb04ac661bcba66"} Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.290893 4998 generic.go:334] "Generic (PLEG): container finished" podID="a556e01d-7f37-46db-ac77-6a681cad9e7e" containerID="9ede5ad8ab91782984292ed4ae96c64b328f5edb554ac5a6084fac7aa87300e7" exitCode=0 Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.291061 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hwftp" event={"ID":"a556e01d-7f37-46db-ac77-6a681cad9e7e","Type":"ContainerDied","Data":"9ede5ad8ab91782984292ed4ae96c64b328f5edb554ac5a6084fac7aa87300e7"} Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.291947 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctpbx" event={"ID":"84e61bdd-8de2-4cd1-93a2-b43d802fbea3","Type":"ContainerStarted","Data":"808664cea365de4f748218334dacb9a55daefc318b8602fc9ee740583ad4b0cd"} Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.294076 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-qvmlb" event={"ID":"e1ff9891-7017-41a2-b7c0-183caa783671","Type":"ContainerStarted","Data":"3e773adab2f6d44884f28afc9643bc13932da091d0ee463ca444d6504b635641"} Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.297040 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" event={"ID":"c2ff7dad-a4da-4243-ab1e-21b47230504d","Type":"ContainerStarted","Data":"c7e0680cfd0e3670e44d7ed7f08e097b2c31d9a9ba81e3d06050a2cf6c2630bc"} Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.297241 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.299700 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wk6qg" event={"ID":"e569f94d-9231-4277-af77-f97c23fcbabe","Type":"ContainerStarted","Data":"ff348035e884090e458dabd0c354866b36eb5d8d913797d8c7692c73166abdc4"} Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.299947 4998 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-2kprn container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.299984 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" podUID="c2ff7dad-a4da-4243-ab1e-21b47230504d" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.302685 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8ft9r" event={"ID":"c71a13b3-a727-4b88-9be8-954a62dc7c86","Type":"ContainerStarted","Data":"e8ccda690ee5766918b7f494c2513f1b79567ebfac02eaec1d9003480bea24cc"} Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.303578 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e6207e94-fd68-464e-9830-895d96a3437c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.303620 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.303645 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e6207e94-fd68-464e-9830-895d96a3437c-registry-tls\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.303730 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a62a491c-a99b-446a-9163-c35451c1e6f2-node-bootstrap-token\") pod \"machine-config-server-gbvbf\" (UID: \"a62a491c-a99b-446a-9163-c35451c1e6f2\") " pod="openshift-machine-config-operator/machine-config-server-gbvbf" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.303871 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e6207e94-fd68-464e-9830-895d96a3437c-registry-certificates\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.303939 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e6207e94-fd68-464e-9830-895d96a3437c-bound-sa-token\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: E1203 16:05:11.303962 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:11.803948858 +0000 UTC m=+90.415649081 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.303981 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e6207e94-fd68-464e-9830-895d96a3437c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.304027 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w56vz\" (UniqueName: \"kubernetes.io/projected/e6207e94-fd68-464e-9830-895d96a3437c-kube-api-access-w56vz\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.304083 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e6207e94-fd68-464e-9830-895d96a3437c-trusted-ca\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.304109 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a62a491c-a99b-446a-9163-c35451c1e6f2-certs\") pod \"machine-config-server-gbvbf\" (UID: \"a62a491c-a99b-446a-9163-c35451c1e6f2\") " pod="openshift-machine-config-operator/machine-config-server-gbvbf" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.304165 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgkcl\" (UniqueName: \"kubernetes.io/projected/a62a491c-a99b-446a-9163-c35451c1e6f2-kube-api-access-pgkcl\") pod \"machine-config-server-gbvbf\" (UID: \"a62a491c-a99b-446a-9163-c35451c1e6f2\") " pod="openshift-machine-config-operator/machine-config-server-gbvbf" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.304741 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jzdp" event={"ID":"20e7b7c5-1480-498e-ab4d-2e335da52dd3","Type":"ContainerStarted","Data":"0bd699f383bf2536af47084dc7eccb757d14f968f385b92d65a2912309d168c6"} Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.307680 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" event={"ID":"56ad3380-4861-49d1-8758-7b1e27f74560","Type":"ContainerStarted","Data":"f6b1e607b8be12fd2922ddb6c59e52604deb2fdad5903c35d87e7586f4f82867"} Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.308892 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-ggw6k" event={"ID":"892e64fc-db6b-4410-9b4c-3ff78cc82074","Type":"ContainerStarted","Data":"509be5dac09d789b7e8dfacafdd4303c6bd938296c865a816fec94201fea5939"} Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.309487 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nsq9q" event={"ID":"09d80f4a-297b-4000-8fcd-c586b5267339","Type":"ContainerStarted","Data":"98adad51b7045a3f8e48b8178fa6bc1786ca5e949c5f9a3b149174631d346283"} Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.355327 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-kk8mb"] Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.405247 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:11 crc kubenswrapper[4998]: E1203 16:05:11.405516 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:11.905493337 +0000 UTC m=+90.517193560 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.405886 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a62a491c-a99b-446a-9163-c35451c1e6f2-node-bootstrap-token\") pod \"machine-config-server-gbvbf\" (UID: \"a62a491c-a99b-446a-9163-c35451c1e6f2\") " pod="openshift-machine-config-operator/machine-config-server-gbvbf" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.406015 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e6207e94-fd68-464e-9830-895d96a3437c-registry-certificates\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.406095 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e6207e94-fd68-464e-9830-895d96a3437c-bound-sa-token\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.406235 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e6207e94-fd68-464e-9830-895d96a3437c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.406391 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w56vz\" (UniqueName: \"kubernetes.io/projected/e6207e94-fd68-464e-9830-895d96a3437c-kube-api-access-w56vz\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.406616 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e6207e94-fd68-464e-9830-895d96a3437c-trusted-ca\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.406672 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a62a491c-a99b-446a-9163-c35451c1e6f2-certs\") pod \"machine-config-server-gbvbf\" (UID: \"a62a491c-a99b-446a-9163-c35451c1e6f2\") " pod="openshift-machine-config-operator/machine-config-server-gbvbf" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.406716 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgkcl\" (UniqueName: \"kubernetes.io/projected/a62a491c-a99b-446a-9163-c35451c1e6f2-kube-api-access-pgkcl\") pod \"machine-config-server-gbvbf\" (UID: \"a62a491c-a99b-446a-9163-c35451c1e6f2\") " pod="openshift-machine-config-operator/machine-config-server-gbvbf" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.409963 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e6207e94-fd68-464e-9830-895d96a3437c-registry-certificates\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.413609 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e6207e94-fd68-464e-9830-895d96a3437c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.414029 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e6207e94-fd68-464e-9830-895d96a3437c-trusted-ca\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.414667 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.414884 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e6207e94-fd68-464e-9830-895d96a3437c-registry-tls\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: E1203 16:05:11.414927 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:11.914916026 +0000 UTC m=+90.526616249 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.422243 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e6207e94-fd68-464e-9830-895d96a3437c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.425973 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a62a491c-a99b-446a-9163-c35451c1e6f2-node-bootstrap-token\") pod \"machine-config-server-gbvbf\" (UID: \"a62a491c-a99b-446a-9163-c35451c1e6f2\") " pod="openshift-machine-config-operator/machine-config-server-gbvbf" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.426555 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a62a491c-a99b-446a-9163-c35451c1e6f2-certs\") pod \"machine-config-server-gbvbf\" (UID: \"a62a491c-a99b-446a-9163-c35451c1e6f2\") " pod="openshift-machine-config-operator/machine-config-server-gbvbf" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.427829 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e6207e94-fd68-464e-9830-895d96a3437c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.438627 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e6207e94-fd68-464e-9830-895d96a3437c-registry-tls\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.452663 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e6207e94-fd68-464e-9830-895d96a3437c-bound-sa-token\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.480301 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w56vz\" (UniqueName: \"kubernetes.io/projected/e6207e94-fd68-464e-9830-895d96a3437c-kube-api-access-w56vz\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: W1203 16:05:11.487120 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a975a4d_12df_4c73_b644_ff281b3c3390.slice/crio-cd425437d010f06ee2d771731323143948468b41c8587af4591bc817ac6dc0d2 WatchSource:0}: Error finding container cd425437d010f06ee2d771731323143948468b41c8587af4591bc817ac6dc0d2: Status 404 returned error can't find the container with id cd425437d010f06ee2d771731323143948468b41c8587af4591bc817ac6dc0d2 Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.498544 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgkcl\" (UniqueName: \"kubernetes.io/projected/a62a491c-a99b-446a-9163-c35451c1e6f2-kube-api-access-pgkcl\") pod \"machine-config-server-gbvbf\" (UID: \"a62a491c-a99b-446a-9163-c35451c1e6f2\") " pod="openshift-machine-config-operator/machine-config-server-gbvbf" Dec 03 16:05:11 crc kubenswrapper[4998]: W1203 16:05:11.507487 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03021110_7732_4dff_a5e3_f481e8b7c0d6.slice/crio-ba405cb10a18a7f717e33aa3807d4a2d514dc108ff7c5e11d02a6883fa993d10 WatchSource:0}: Error finding container ba405cb10a18a7f717e33aa3807d4a2d514dc108ff7c5e11d02a6883fa993d10: Status 404 returned error can't find the container with id ba405cb10a18a7f717e33aa3807d4a2d514dc108ff7c5e11d02a6883fa993d10 Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.524878 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vns7l"] Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.528235 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:11 crc kubenswrapper[4998]: E1203 16:05:11.528707 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:12.028689836 +0000 UTC m=+90.640390059 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.535374 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-ppw5v"] Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.601919 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-gbvbf" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.624869 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9"] Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.632331 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: E1203 16:05:11.632838 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:12.132822282 +0000 UTC m=+90.744522505 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.656645 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5wc9c"] Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.656691 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-r2gbh"] Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.664278 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412960-6vr8n"] Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.710697 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-jjwjb"] Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.716166 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" podStartSLOduration=71.716147228 podStartE2EDuration="1m11.716147228s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:11.716009445 +0000 UTC m=+90.327709668" watchObservedRunningTime="2025-12-03 16:05:11.716147228 +0000 UTC m=+90.327847451" Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.733508 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:11 crc kubenswrapper[4998]: E1203 16:05:11.733606 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:12.233587254 +0000 UTC m=+90.845287477 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.733762 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: E1203 16:05:11.734575 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:12.234558196 +0000 UTC m=+90.846258419 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.784570 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sn2r4"] Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.834825 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:11 crc kubenswrapper[4998]: E1203 16:05:11.835108 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:12.335091723 +0000 UTC m=+90.946791946 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.854684 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-4rq7l"] Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.856732 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-54vrz"] Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.861214 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wfm8s"] Dec 03 16:05:11 crc kubenswrapper[4998]: W1203 16:05:11.873799 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22c37be4_5e07_4638_aa01_04ee06c463d7.slice/crio-b76b46a674643bacf5eb206e79cf7e8f2de42406e8379b6ade876971b37034b5 WatchSource:0}: Error finding container b76b46a674643bacf5eb206e79cf7e8f2de42406e8379b6ade876971b37034b5: Status 404 returned error can't find the container with id b76b46a674643bacf5eb206e79cf7e8f2de42406e8379b6ade876971b37034b5 Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.898254 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-wsmkz"] Dec 03 16:05:11 crc kubenswrapper[4998]: I1203 16:05:11.940860 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:11 crc kubenswrapper[4998]: E1203 16:05:11.941480 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:12.441437878 +0000 UTC m=+91.053138101 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.042309 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:12 crc kubenswrapper[4998]: E1203 16:05:12.042577 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:12.542563428 +0000 UTC m=+91.154263651 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.150309 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:12 crc kubenswrapper[4998]: E1203 16:05:12.150573 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:12.65056356 +0000 UTC m=+91.262263783 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.212461 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rrj4k"] Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.250856 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:12 crc kubenswrapper[4998]: E1203 16:05:12.251282 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:12.751266441 +0000 UTC m=+91.362966664 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.255803 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-5c2lp"] Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.261979 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-k92zl"] Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.323090 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8lmkd" event={"ID":"7a975a4d-12df-4c73-b644-ff281b3c3390","Type":"ContainerStarted","Data":"cd425437d010f06ee2d771731323143948468b41c8587af4591bc817ac6dc0d2"} Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.326631 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-4rq7l" event={"ID":"57c54ccc-a846-4c00-8318-5d5331c70e70","Type":"ContainerStarted","Data":"8f8b34772c44fb5a243615f344985052bffac2fea1dec7bcdf7298ade35f7ab3"} Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.339544 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5wc9c" event={"ID":"68597ef9-34cf-45c4-86c7-153f9f4d9363","Type":"ContainerStarted","Data":"7bf1433afa62914d594c0c0a759d54efcfc06787d7aa5134e555a67b23bcf527"} Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.341537 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9cdmt" event={"ID":"94922a63-40d9-4944-9d95-f860967ea5b3","Type":"ContainerStarted","Data":"69c4c927d59c652324d12d14f2b26b932dafa4ec19c5ffc13c90edc4b9fd5f8a"} Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.342421 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9" event={"ID":"6024dcd8-96d6-4c42-9670-42e3608cd7ee","Type":"ContainerStarted","Data":"a1ef71f0e0b935d309143ccff1a8427f30cc84e052bf2b5d70fe371f7150b2b7"} Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.347862 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sn2r4" event={"ID":"12d866d2-1002-4db4-9832-59b5fea5062e","Type":"ContainerStarted","Data":"da2b906308e805b3c52c69e823e8a0930578d5a3a1ef76becbbb4aa85a951eb6"} Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.352360 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:12 crc kubenswrapper[4998]: E1203 16:05:12.352681 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:12.852670737 +0000 UTC m=+91.464370960 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.406231 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-ppw5v" event={"ID":"2e58b61f-82bf-42c2-a664-cd3fcbd0fc03","Type":"ContainerStarted","Data":"cb69050322f9fa2684b3b11f5ab949d95c00585967527dfa0e5295db8bac9bb3"} Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.425418 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-ggw6k" event={"ID":"892e64fc-db6b-4410-9b4c-3ff78cc82074","Type":"ContainerStarted","Data":"0ff9df47dfea58693912c7f2f623efd31e8496a8f3cae67bbbaf3c9ef5cc42e9"} Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.439744 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-qvmlb" event={"ID":"e1ff9891-7017-41a2-b7c0-183caa783671","Type":"ContainerStarted","Data":"ee93e469285639d372ccba535873ef555396811b0b403aad5961135a1afcfaf3"} Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.440803 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-qvmlb" Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.459827 4998 patch_prober.go:28] interesting pod/console-operator-58897d9998-qvmlb container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.459873 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-qvmlb" podUID="e1ff9891-7017-41a2-b7c0-183caa783671" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.460655 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:12 crc kubenswrapper[4998]: E1203 16:05:12.460994 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:12.960981746 +0000 UTC m=+91.572681969 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.461079 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-54vrz" event={"ID":"0a1b6dae-ff3d-4bfb-a811-98866c9c2b59","Type":"ContainerStarted","Data":"69ea0315436fca2ae40c0454f224ff1072ef29143353a536cd3c1ec49a05c76c"} Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.474972 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wk6qg" event={"ID":"e569f94d-9231-4277-af77-f97c23fcbabe","Type":"ContainerStarted","Data":"6a86e482ca2d241b53dbddedcf7a07d3058daafb39a20018a12c23bc8723ecf5"} Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.532522 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8ft9r" event={"ID":"c71a13b3-a727-4b88-9be8-954a62dc7c86","Type":"ContainerStarted","Data":"32ec9e98de705e1bc960ab1173db07dacbb4e01cb983df15dfc6b19fd4c05aea"} Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.549549 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-wsmkz" event={"ID":"a741e843-b85d-4250-844a-b03534dc47d2","Type":"ContainerStarted","Data":"9dd7f29678adb6e434105bfa7fa987807ab6c9653ea77593f8d55b9ee374234a"} Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.551029 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-r2gbh" event={"ID":"22c37be4-5e07-4638-aa01-04ee06c463d7","Type":"ContainerStarted","Data":"b76b46a674643bacf5eb206e79cf7e8f2de42406e8379b6ade876971b37034b5"} Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.562083 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:12 crc kubenswrapper[4998]: E1203 16:05:12.563064 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:13.063043917 +0000 UTC m=+91.674744220 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.563960 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-hnvvp" event={"ID":"cbae1035-ec59-4f24-884d-809c7e7bb685","Type":"ContainerStarted","Data":"a6f02cf897e1b3fb7422cb8588cbb5bfb3950c8cbf17a9b295c5fd1d0eefb699"} Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.566054 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" event={"ID":"77eb0bc5-1a4a-48ae-98cc-f26a450af04a","Type":"ContainerStarted","Data":"2282d2e121f6ec2e4b25964d414e9c0909893d1536141b407e57de4d9e6333eb"} Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.583842 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-b9f5r"] Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.584434 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-dc55h"] Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.589314 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jjwjb" event={"ID":"8c88814f-7a99-4878-9517-55cbecd72ff0","Type":"ContainerStarted","Data":"2bfcedaa8ee354e385cb307fac6d75725aa2f6fff9a3c39a9739fe0287b82adb"} Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.612514 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nsq9q" event={"ID":"09d80f4a-297b-4000-8fcd-c586b5267339","Type":"ContainerStarted","Data":"716b80b34caeb9149d1e3a3094bd597f6ae50172a0f50774c5630258b33ea83d"} Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.614353 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wfm8s" event={"ID":"31bfedfc-39de-4bf0-b049-e41c577ac848","Type":"ContainerStarted","Data":"cbb12db7a2d38f0befa876518455537a80550526c225e701a041c0ec56763630"} Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.619627 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kk8mb" event={"ID":"03021110-7732-4dff-a5e3-f481e8b7c0d6","Type":"ContainerStarted","Data":"ba405cb10a18a7f717e33aa3807d4a2d514dc108ff7c5e11d02a6883fa993d10"} Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.622169 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xd7m2"] Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.642201 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-gbvbf" event={"ID":"a62a491c-a99b-446a-9163-c35451c1e6f2","Type":"ContainerStarted","Data":"3de398a47185353e323319dc6e5664c633e6980c29fb22615af35ec7c1d509cf"} Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.646086 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kg4lw" event={"ID":"ac8c6176-0450-4050-9094-003067cd40b1","Type":"ContainerDied","Data":"3f501819e2212bc90fcae42b3c1688daa04a5fffa9907d0065c5d6c0c5a8599a"} Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.652740 4998 generic.go:334] "Generic (PLEG): container finished" podID="ac8c6176-0450-4050-9094-003067cd40b1" containerID="3f501819e2212bc90fcae42b3c1688daa04a5fffa9907d0065c5d6c0c5a8599a" exitCode=0 Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.661020 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctpbx" event={"ID":"84e61bdd-8de2-4cd1-93a2-b43d802fbea3","Type":"ContainerStarted","Data":"b5de961ad8d3fe2523cf0577eb53aa02e1cf33a6b4bede6d590ce5d8033f0f48"} Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.662568 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:12 crc kubenswrapper[4998]: E1203 16:05:12.663447 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:13.163432161 +0000 UTC m=+91.775132384 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.669106 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-6vr8n" event={"ID":"a6e6bbd8-c854-4708-b9b1-40884d1be6c7","Type":"ContainerStarted","Data":"a26607d85381891eef17c236245f6e92b5a22cf8e6587d7a928a037fedf2b362"} Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.686220 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" event={"ID":"56ad3380-4861-49d1-8758-7b1e27f74560","Type":"ContainerStarted","Data":"4cfc40eef4ee0dce61eb0b953949c62810ea4c11b954526d5e8450ec31a260fa"} Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.686292 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.699597 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.769670 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:12 crc kubenswrapper[4998]: E1203 16:05:12.772676 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:13.27266093 +0000 UTC m=+91.884361153 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.777787 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-bhj2p"] Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.787824 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fnxt7"] Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.789405 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-v8c75"] Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.872140 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:12 crc kubenswrapper[4998]: E1203 16:05:12.872518 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:13.372503422 +0000 UTC m=+91.984203645 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:12 crc kubenswrapper[4998]: W1203 16:05:12.936877 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ce38b01_0835_4a45_a262_f584fe51bd62.slice/crio-a242d8a1dca7510c5af8579fdc6024da150a7c9ac9b040e631f41807d7441dc2 WatchSource:0}: Error finding container a242d8a1dca7510c5af8579fdc6024da150a7c9ac9b040e631f41807d7441dc2: Status 404 returned error can't find the container with id a242d8a1dca7510c5af8579fdc6024da150a7c9ac9b040e631f41807d7441dc2 Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.960416 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-5mthf"] Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.972043 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l9b5"] Dec 03 16:05:12 crc kubenswrapper[4998]: I1203 16:05:12.978987 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:12 crc kubenswrapper[4998]: E1203 16:05:12.979259 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:13.479249517 +0000 UTC m=+92.090949740 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.005070 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jmdsz"] Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.080212 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:13 crc kubenswrapper[4998]: E1203 16:05:13.080565 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:13.580549 +0000 UTC m=+92.192249223 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.085707 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.136461 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zh6tw"] Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.182364 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:13 crc kubenswrapper[4998]: E1203 16:05:13.185097 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:13.685075916 +0000 UTC m=+92.296776129 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:13 crc kubenswrapper[4998]: W1203 16:05:13.202042 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7274d4ec_5476_46c7_ab6a_f4ed0f4b294e.slice/crio-aca34701052c4ada9ac8af7879be9775a7c09844fc7ad58c0a44629d7f8b834f WatchSource:0}: Error finding container aca34701052c4ada9ac8af7879be9775a7c09844fc7ad58c0a44629d7f8b834f: Status 404 returned error can't find the container with id aca34701052c4ada9ac8af7879be9775a7c09844fc7ad58c0a44629d7f8b834f Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.283719 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:13 crc kubenswrapper[4998]: E1203 16:05:13.284249 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:13.784001837 +0000 UTC m=+92.395702060 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.319782 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" podStartSLOduration=73.319764429 podStartE2EDuration="1m13.319764429s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:13.317260904 +0000 UTC m=+91.928961137" watchObservedRunningTime="2025-12-03 16:05:13.319764429 +0000 UTC m=+91.931464652" Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.387528 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:13 crc kubenswrapper[4998]: E1203 16:05:13.387893 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:13.887876338 +0000 UTC m=+92.499576561 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.471330 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-qvmlb" podStartSLOduration=73.471312596 podStartE2EDuration="1m13.471312596s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:13.470201592 +0000 UTC m=+92.081901815" watchObservedRunningTime="2025-12-03 16:05:13.471312596 +0000 UTC m=+92.083012819" Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.488452 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:13 crc kubenswrapper[4998]: E1203 16:05:13.489329 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:13.989314605 +0000 UTC m=+92.601014828 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.590736 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:13 crc kubenswrapper[4998]: E1203 16:05:13.591160 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:14.091148891 +0000 UTC m=+92.702849104 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.624154 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nsq9q" podStartSLOduration=73.624135561 podStartE2EDuration="1m13.624135561s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:13.58569234 +0000 UTC m=+92.197392563" watchObservedRunningTime="2025-12-03 16:05:13.624135561 +0000 UTC m=+92.235835784" Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.649054 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctpbx" podStartSLOduration=73.649040553 podStartE2EDuration="1m13.649040553s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:13.647459258 +0000 UTC m=+92.259159481" watchObservedRunningTime="2025-12-03 16:05:13.649040553 +0000 UTC m=+92.260740776" Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.692216 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:13 crc kubenswrapper[4998]: E1203 16:05:13.692369 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:14.192353052 +0000 UTC m=+92.804053275 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.692785 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:13 crc kubenswrapper[4998]: E1203 16:05:13.693117 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:14.193109699 +0000 UTC m=+92.804809912 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.720298 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-ppw5v" event={"ID":"2e58b61f-82bf-42c2-a664-cd3fcbd0fc03","Type":"ContainerStarted","Data":"91b7fb1555645e1a728d15afcdded38242a9587b5c32561ea384c847aab16f6e"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.720772 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-ppw5v" Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.721779 4998 patch_prober.go:28] interesting pod/downloads-7954f5f757-ppw5v container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.721823 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ppw5v" podUID="2e58b61f-82bf-42c2-a664-cd3fcbd0fc03" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.31:8080/\": dial tcp 10.217.0.31:8080: connect: connection refused" Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.725077 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hwftp" event={"ID":"a556e01d-7f37-46db-ac77-6a681cad9e7e","Type":"ContainerStarted","Data":"86fe062a5edcbbc290535bea424c5fd9117e594ca8b045af16e9f7df0546653c"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.728194 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l9b5" event={"ID":"02311a5a-a912-4cce-b076-d752266f0c95","Type":"ContainerStarted","Data":"3bff7cdae3f554e05b1c840f66898b7d9a54c696cbdb21ec415629a0ca33f6a0"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.730294 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fnxt7" event={"ID":"8c715a96-664c-4b26-8d5a-523a0fb4a17c","Type":"ContainerStarted","Data":"bf682da7e63db18e5efb3089e9cf4cba51cc142a82e7c8e5581fb1b95efe6c4e"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.736387 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jmdsz" event={"ID":"7274d4ec-5476-46c7-ab6a-f4ed0f4b294e","Type":"ContainerStarted","Data":"aca34701052c4ada9ac8af7879be9775a7c09844fc7ad58c0a44629d7f8b834f"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.743963 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xd7m2" event={"ID":"9a6a4ae3-f219-4a1b-bfbe-f9f829945a8a","Type":"ContainerStarted","Data":"ddf5f73b85079266c89f86a2d78602da7bd2b8e91d184185c1bd47dd9e1dc155"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.746367 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zh6tw" event={"ID":"bff7045b-cac0-438c-b6a6-769147c3f108","Type":"ContainerStarted","Data":"6467218dd0e4d430655653d66af74624fb7b5bb5b0638c807c0f63c7c3c2a040"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.751115 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wk6qg" podStartSLOduration=73.751091333 podStartE2EDuration="1m13.751091333s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:13.748775712 +0000 UTC m=+92.360475955" watchObservedRunningTime="2025-12-03 16:05:13.751091333 +0000 UTC m=+92.362791546" Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.758338 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-ggw6k" event={"ID":"892e64fc-db6b-4410-9b4c-3ff78cc82074","Type":"ContainerStarted","Data":"b8485f66f90133b20a449f547e676386d0f9a7504b68388b2f34cd86bac0eafb"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.776271 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jjwjb" event={"ID":"8c88814f-7a99-4878-9517-55cbecd72ff0","Type":"ContainerStarted","Data":"5c882aa8466c5ce2a035c09a0b9da85e6cc37336aa5fd27affe5843bfc7d8934"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.787115 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8ft9r" event={"ID":"c71a13b3-a727-4b88-9be8-954a62dc7c86","Type":"ContainerStarted","Data":"9463b7c1d9d8cea302e18ed0ad0673f7978f68788159537e9ec2f446081c2316"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.793921 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.793953 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rrj4k" event={"ID":"a29a797b-0ffe-4e7c-b4c4-d3869fb694fe","Type":"ContainerStarted","Data":"001e418c2eb8d0abe6b66f649a93fe079dd57e89a31921870e3e6416d10ab061"} Dec 03 16:05:13 crc kubenswrapper[4998]: E1203 16:05:13.794224 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:14.294202138 +0000 UTC m=+92.905902371 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.795072 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5mthf" event={"ID":"e7a61bf3-1760-4910-9561-000d68ca9708","Type":"ContainerStarted","Data":"e4efcc87cea0ac15dd19d95b2f0536bef4ea9060bd559a00bbb718983645c605"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.803010 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-5c2lp" event={"ID":"235934fb-ec04-4784-b30c-5c80dfa325ad","Type":"ContainerStarted","Data":"8290f3e2e611a7d9ccd0fe22a126be82a0d55bbdfe42bf7c22364cc61424b582"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.807240 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-54vrz" event={"ID":"0a1b6dae-ff3d-4bfb-a811-98866c9c2b59","Type":"ContainerStarted","Data":"ffb871cad33e4b700cbadaaf14a00937d12703e46d86225839a7cc150628c5dd"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.813394 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jzdp" event={"ID":"20e7b7c5-1480-498e-ab4d-2e335da52dd3","Type":"ContainerStarted","Data":"b7b0da26ffdaf52f9605cef99c25bffc7104dbca5de164d2ef52822fe85cb92b"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.821053 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" event={"ID":"b46224fb-c151-4ba9-87c7-85e0ed6e11bc","Type":"ContainerStarted","Data":"3a87526a8e1b228796b38d831050562b8e47e2061720ad4957360ec100ab746a"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.849403 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9" event={"ID":"6024dcd8-96d6-4c42-9670-42e3608cd7ee","Type":"ContainerStarted","Data":"3eee04750fd3bb67978aba152c91e073fc417746f33a9b9b563171b94743b1a2"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.850405 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9" Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.871701 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-hnvvp" event={"ID":"cbae1035-ec59-4f24-884d-809c7e7bb685","Type":"ContainerStarted","Data":"9106ff039211608da1026b9e0ddfe92a3a31a95225d4b35b07d6d95e09a4252f"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.880015 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" event={"ID":"77eb0bc5-1a4a-48ae-98cc-f26a450af04a","Type":"ContainerStarted","Data":"73bc2075fc9c95fa307e7d27df36b0a6e1af14adb0cb95dd5919729db2b9ae03"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.892384 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kk8mb" event={"ID":"03021110-7732-4dff-a5e3-f481e8b7c0d6","Type":"ContainerStarted","Data":"86442755efbc807979332830ff8cf6a8cdce2a791eb5a81ea3f7aac061669b6d"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.893488 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-ggw6k" podStartSLOduration=73.893477247 podStartE2EDuration="1m13.893477247s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:13.89223728 +0000 UTC m=+92.503937503" watchObservedRunningTime="2025-12-03 16:05:13.893477247 +0000 UTC m=+92.505177470" Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.895431 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:13 crc kubenswrapper[4998]: E1203 16:05:13.901838 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:14.401823172 +0000 UTC m=+93.013523395 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.904194 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dc55h" event={"ID":"f089be95-2900-436d-b1f4-627095e2d563","Type":"ContainerStarted","Data":"78a22c35ea462c10cccd980f3ff83c3b48a60041ef65cf34aed4a42a26e259f4"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.912941 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-v8c75" event={"ID":"75498a9e-2bcc-4e76-a528-16df0f6d8ae7","Type":"ContainerStarted","Data":"9a7b15535d7224423ef4aebf4769b4cff5a4fe3383225079c68fae6176daf0ca"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.925020 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wfm8s" event={"ID":"31bfedfc-39de-4bf0-b049-e41c577ac848","Type":"ContainerStarted","Data":"ba0a5e8d3fa5045397e9452ea2cfb9dab44d47bb2d15c2582528dfaa7fb3907c"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.925881 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wfm8s" Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.927658 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8ft9r" podStartSLOduration=73.927648174 podStartE2EDuration="1m13.927648174s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:13.926169911 +0000 UTC m=+92.537870134" watchObservedRunningTime="2025-12-03 16:05:13.927648174 +0000 UTC m=+92.539348387" Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.928789 4998 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-wfm8s container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.928852 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wfm8s" podUID="31bfedfc-39de-4bf0-b049-e41c577ac848" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.950777 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-b9f5r" event={"ID":"6ce38b01-0835-4a45-a262-f584fe51bd62","Type":"ContainerStarted","Data":"a242d8a1dca7510c5af8579fdc6024da150a7c9ac9b040e631f41807d7441dc2"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.965391 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8lmkd" event={"ID":"7a975a4d-12df-4c73-b644-ff281b3c3390","Type":"ContainerStarted","Data":"dc694ec59af86a432840a3a07a0aae41fbfab2f5dcbb7e2bf1841284438f2a35"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.975534 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bhj2p" event={"ID":"ef0bf68e-f7d5-43a2-890c-10ffba3e6238","Type":"ContainerStarted","Data":"4f95100875f65ebd8b39088c07acaf493273b09e2e45a8e8cd176cc1bf3fedf7"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.978137 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-k92zl" event={"ID":"817c8b8a-cab9-474a-ac82-30b55677c2f4","Type":"ContainerStarted","Data":"d6077eb3980c756f0fb581c0538291b5f513762ce429fabef9f058a63aa742dc"} Dec 03 16:05:13 crc kubenswrapper[4998]: I1203 16:05:13.997811 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:13 crc kubenswrapper[4998]: E1203 16:05:13.998932 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:14.498905492 +0000 UTC m=+93.110605715 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.000253 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-qvmlb" Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.012143 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-54vrz" podStartSLOduration=74.012124865 podStartE2EDuration="1m14.012124865s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:14.007295758 +0000 UTC m=+92.618995981" watchObservedRunningTime="2025-12-03 16:05:14.012124865 +0000 UTC m=+92.623825088" Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.012543 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-ppw5v" podStartSLOduration=74.012539334 podStartE2EDuration="1m14.012539334s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:13.958367105 +0000 UTC m=+92.570067328" watchObservedRunningTime="2025-12-03 16:05:14.012539334 +0000 UTC m=+92.624239557" Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.089491 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" podStartSLOduration=73.089476429 podStartE2EDuration="1m13.089476429s" podCreationTimestamp="2025-12-03 16:04:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:14.087999736 +0000 UTC m=+92.699699959" watchObservedRunningTime="2025-12-03 16:05:14.089476429 +0000 UTC m=+92.701176652" Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.101169 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:14 crc kubenswrapper[4998]: E1203 16:05:14.112777 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:14.612747414 +0000 UTC m=+93.224447637 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.137845 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-hnvvp" Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.165074 4998 patch_prober.go:28] interesting pod/router-default-5444994796-hnvvp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 16:05:14 crc kubenswrapper[4998]: [-]has-synced failed: reason withheld Dec 03 16:05:14 crc kubenswrapper[4998]: [+]process-running ok Dec 03 16:05:14 crc kubenswrapper[4998]: healthz check failed Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.165122 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hnvvp" podUID="cbae1035-ec59-4f24-884d-809c7e7bb685" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.183321 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9" podStartSLOduration=73.183305337 podStartE2EDuration="1m13.183305337s" podCreationTimestamp="2025-12-03 16:04:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:14.127421159 +0000 UTC m=+92.739121382" watchObservedRunningTime="2025-12-03 16:05:14.183305337 +0000 UTC m=+92.795005560" Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.184942 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-kk8mb" podStartSLOduration=74.184936803 podStartE2EDuration="1m14.184936803s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:14.182398677 +0000 UTC m=+92.794098900" watchObservedRunningTime="2025-12-03 16:05:14.184936803 +0000 UTC m=+92.796637026" Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.205603 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:14 crc kubenswrapper[4998]: E1203 16:05:14.206001 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:14.705986789 +0000 UTC m=+93.317687002 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.206820 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6jzdp" podStartSLOduration=75.206801027 podStartE2EDuration="1m15.206801027s" podCreationTimestamp="2025-12-03 16:03:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:14.205227563 +0000 UTC m=+92.816927776" watchObservedRunningTime="2025-12-03 16:05:14.206801027 +0000 UTC m=+92.818501250" Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.272501 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-vns7l" podStartSLOduration=74.272482982 podStartE2EDuration="1m14.272482982s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:14.270267083 +0000 UTC m=+92.881967306" watchObservedRunningTime="2025-12-03 16:05:14.272482982 +0000 UTC m=+92.884183205" Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.312785 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:14 crc kubenswrapper[4998]: E1203 16:05:14.313132 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:14.813117992 +0000 UTC m=+93.424818215 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.326099 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-hnvvp" podStartSLOduration=74.326082739 podStartE2EDuration="1m14.326082739s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:14.324918764 +0000 UTC m=+92.936618987" watchObservedRunningTime="2025-12-03 16:05:14.326082739 +0000 UTC m=+92.937782962" Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.377027 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wfm8s" podStartSLOduration=73.377012948 podStartE2EDuration="1m13.377012948s" podCreationTimestamp="2025-12-03 16:04:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:14.374198515 +0000 UTC m=+92.985898738" watchObservedRunningTime="2025-12-03 16:05:14.377012948 +0000 UTC m=+92.988713171" Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.378855 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.378903 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.413704 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8lmkd" podStartSLOduration=74.41368976 podStartE2EDuration="1m14.41368976s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:14.411786888 +0000 UTC m=+93.023487111" watchObservedRunningTime="2025-12-03 16:05:14.41368976 +0000 UTC m=+93.025389983" Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.415074 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:14 crc kubenswrapper[4998]: E1203 16:05:14.415441 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:14.915427519 +0000 UTC m=+93.527127742 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.415984 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.517099 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:14 crc kubenswrapper[4998]: E1203 16:05:14.517778 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:15.017764984 +0000 UTC m=+93.629465197 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.543048 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9" Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.618369 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:14 crc kubenswrapper[4998]: E1203 16:05:14.618672 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:15.118655329 +0000 UTC m=+93.730355552 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.720255 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:14 crc kubenswrapper[4998]: E1203 16:05:14.720895 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:15.220882544 +0000 UTC m=+93.832582767 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.822170 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:14 crc kubenswrapper[4998]: E1203 16:05:14.822863 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:15.322846422 +0000 UTC m=+93.934546645 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:14 crc kubenswrapper[4998]: I1203 16:05:14.924297 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:14 crc kubenswrapper[4998]: E1203 16:05:14.924827 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:15.424815151 +0000 UTC m=+94.036515374 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.025360 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:15 crc kubenswrapper[4998]: E1203 16:05:15.025656 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:15.525641774 +0000 UTC m=+94.137341987 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.030707 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9cdmt" event={"ID":"94922a63-40d9-4944-9d95-f860967ea5b3","Type":"ContainerStarted","Data":"3de909bf8474873b2f6904f3a61fd1bd865704a696d86cddd9cb8920edb29628"} Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.036020 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sn2r4" event={"ID":"12d866d2-1002-4db4-9832-59b5fea5062e","Type":"ContainerStarted","Data":"e2e153e32de6bf4c7a9f4c05078c48d712bdcae5472600b61200fdaea4150fea"} Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.042230 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-wsmkz" event={"ID":"a741e843-b85d-4250-844a-b03534dc47d2","Type":"ContainerStarted","Data":"8b9c8e769806a98b633f73fa4ee399ccfe80315879a97c77f730a76950973cdb"} Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.042272 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-wsmkz" event={"ID":"a741e843-b85d-4250-844a-b03534dc47d2","Type":"ContainerStarted","Data":"8ee14de41c1c50f57ce5eabedcc7e5c32436fc9eabc89c7dc28e5535a91a55e1"} Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.042992 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-wsmkz" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.062609 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-9cdmt" podStartSLOduration=76.062590573 podStartE2EDuration="1m16.062590573s" podCreationTimestamp="2025-12-03 16:03:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:15.061338245 +0000 UTC m=+93.673038468" watchObservedRunningTime="2025-12-03 16:05:15.062590573 +0000 UTC m=+93.674290796" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.079294 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-hwftp" event={"ID":"a556e01d-7f37-46db-ac77-6a681cad9e7e","Type":"ContainerStarted","Data":"1138f2a0a92e10a693a4d03511a6a30ce7dd61271136724c9ea127ca0d10634c"} Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.087206 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sn2r4" podStartSLOduration=75.087188548 podStartE2EDuration="1m15.087188548s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:15.085207224 +0000 UTC m=+93.696907447" watchObservedRunningTime="2025-12-03 16:05:15.087188548 +0000 UTC m=+93.698888771" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.099317 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kg4lw" event={"ID":"ac8c6176-0450-4050-9094-003067cd40b1","Type":"ContainerStarted","Data":"8b1898f58d83732e0bf21378620070a10b4ab72cff33e4d4109df01d512c0634"} Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.099966 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kg4lw" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.118986 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l9b5" event={"ID":"02311a5a-a912-4cce-b076-d752266f0c95","Type":"ContainerStarted","Data":"7afdfe1fd79355770afdc29ab1692d258a26c556207462a0493f397b22222429"} Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.129167 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:15 crc kubenswrapper[4998]: E1203 16:05:15.130841 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:15.630823974 +0000 UTC m=+94.242524287 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.134687 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-b9f5r" event={"ID":"6ce38b01-0835-4a45-a262-f584fe51bd62","Type":"ContainerStarted","Data":"da0ec523336c7745df82493abfcb62b9fac9159d27c7818477c63e2fd684c457"} Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.143122 4998 patch_prober.go:28] interesting pod/router-default-5444994796-hnvvp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 16:05:15 crc kubenswrapper[4998]: [-]has-synced failed: reason withheld Dec 03 16:05:15 crc kubenswrapper[4998]: [+]process-running ok Dec 03 16:05:15 crc kubenswrapper[4998]: healthz check failed Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.143174 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hnvvp" podUID="cbae1035-ec59-4f24-884d-809c7e7bb685" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.145117 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jmdsz" event={"ID":"7274d4ec-5476-46c7-ab6a-f4ed0f4b294e","Type":"ContainerStarted","Data":"8e49a397ea71f9c3e92eddccc73ecf3fe34e612458877de167dba259068d4888"} Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.150828 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8l9b5" podStartSLOduration=75.150812077 podStartE2EDuration="1m15.150812077s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:15.149063098 +0000 UTC m=+93.760763321" watchObservedRunningTime="2025-12-03 16:05:15.150812077 +0000 UTC m=+93.762512300" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.151856 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-wsmkz" podStartSLOduration=7.15184942 podStartE2EDuration="7.15184942s" podCreationTimestamp="2025-12-03 16:05:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:15.127970991 +0000 UTC m=+93.739671214" watchObservedRunningTime="2025-12-03 16:05:15.15184942 +0000 UTC m=+93.763549643" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.167045 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-6vr8n" event={"ID":"a6e6bbd8-c854-4708-b9b1-40884d1be6c7","Type":"ContainerStarted","Data":"481b1802323e8456ca60f05e1a85214ce7a71c0cba6ebdbe39be6ec00757d272"} Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.176059 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rrj4k" event={"ID":"a29a797b-0ffe-4e7c-b4c4-d3869fb694fe","Type":"ContainerStarted","Data":"0e9f0d1e29bb8d0d106a65614396449c1d4ce4680cc3dc3210660b45fb84ad17"} Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.182008 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-hwftp" podStartSLOduration=75.181994398 podStartE2EDuration="1m15.181994398s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:15.180716399 +0000 UTC m=+93.792416622" watchObservedRunningTime="2025-12-03 16:05:15.181994398 +0000 UTC m=+93.793694621" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.186399 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-v8c75" event={"ID":"75498a9e-2bcc-4e76-a528-16df0f6d8ae7","Type":"ContainerStarted","Data":"8ef7697ad560adc5a47a7753fa5777931527343e0cb7de59b3e4ce268b33a0e0"} Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.192747 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fnxt7" event={"ID":"8c715a96-664c-4b26-8d5a-523a0fb4a17c","Type":"ContainerStarted","Data":"0865731cfb451c650cd07d5dcb67765501d216049957bb9aa334a71a6fd665e8"} Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.192800 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-fnxt7" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.197187 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dc55h" event={"ID":"f089be95-2900-436d-b1f4-627095e2d563","Type":"ContainerStarted","Data":"06fc1f33e4b0620f180cd55be174f687f8c1091e89edfb48bcf99a57ba3c48be"} Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.198595 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xd7m2" event={"ID":"9a6a4ae3-f219-4a1b-bfbe-f9f829945a8a","Type":"ContainerStarted","Data":"1e48ef363e103295220427b9560bf0cf8918994a8fe8765d8a60bf13e426c4da"} Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.199241 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xd7m2" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.199708 4998 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-fnxt7 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.199835 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-fnxt7" podUID="8c715a96-664c-4b26-8d5a-523a0fb4a17c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.204895 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-4rq7l" event={"ID":"57c54ccc-a846-4c00-8318-5d5331c70e70","Type":"ContainerStarted","Data":"35c676d4f3e9ba21f01a0548e9a41364244bc062019dbb2f897fc3d1ec97c40e"} Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.204942 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kg4lw" podStartSLOduration=75.204930496 podStartE2EDuration="1m15.204930496s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:15.199844673 +0000 UTC m=+93.811544896" watchObservedRunningTime="2025-12-03 16:05:15.204930496 +0000 UTC m=+93.816630719" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.208363 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5wc9c" event={"ID":"68597ef9-34cf-45c4-86c7-153f9f4d9363","Type":"ContainerStarted","Data":"69a41312b6c6351a968bcce884a15397b51834352eacf052d32f3f3741d9a4d0"} Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.211008 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-gbvbf" event={"ID":"a62a491c-a99b-446a-9163-c35451c1e6f2","Type":"ContainerStarted","Data":"9f1ec9d81bae711c5bc2caf638ee6591d6480d746cdbbe4538b784d1d615b23c"} Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.212327 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-r2gbh" event={"ID":"22c37be4-5e07-4638-aa01-04ee06c463d7","Type":"ContainerStarted","Data":"4c4f616aa109683e771348a837ea7f0d517dbd5a8b00965c532ca15ade5e8d45"} Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.213380 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jjwjb" event={"ID":"8c88814f-7a99-4878-9517-55cbecd72ff0","Type":"ContainerStarted","Data":"99d00e1a3dd3d86924ad555cc4415d1a35db4411d28ed35cc3b572f29514c8f9"} Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.214548 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zh6tw" event={"ID":"bff7045b-cac0-438c-b6a6-769147c3f108","Type":"ContainerStarted","Data":"97728e4a3bb2f35b5ea4bd94edc682eacb8da15c67b05e8ec83f04d9a9718854"} Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.217083 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zh6tw" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.217134 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-k92zl" event={"ID":"817c8b8a-cab9-474a-ac82-30b55677c2f4","Type":"ContainerStarted","Data":"cdff9bd18edeaf919a9a34cd00bc9410770798e5daaaae8a5467b1a9f62703eb"} Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.220870 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xd7m2" podStartSLOduration=74.220857378 podStartE2EDuration="1m14.220857378s" podCreationTimestamp="2025-12-03 16:04:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:15.22002333 +0000 UTC m=+93.831723553" watchObservedRunningTime="2025-12-03 16:05:15.220857378 +0000 UTC m=+93.832557601" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.226328 4998 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-zh6tw container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:5443/healthz\": dial tcp 10.217.0.24:5443: connect: connection refused" start-of-body= Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.226400 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zh6tw" podUID="bff7045b-cac0-438c-b6a6-769147c3f108" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.24:5443/healthz\": dial tcp 10.217.0.24:5443: connect: connection refused" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.227780 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-5c2lp" event={"ID":"235934fb-ec04-4784-b30c-5c80dfa325ad","Type":"ContainerStarted","Data":"1e3aecbc497c8d5e813f2ceff875a7e55bca2807c590d28c7622941c32c3e6da"} Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.230540 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bhj2p" event={"ID":"ef0bf68e-f7d5-43a2-890c-10ffba3e6238","Type":"ContainerStarted","Data":"4527eafa6917ad528638ff4a73359784e0b67d71ed4071f59043c76c14962e7e"} Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.231198 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.231858 4998 patch_prober.go:28] interesting pod/downloads-7954f5f757-ppw5v container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.231898 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ppw5v" podUID="2e58b61f-82bf-42c2-a664-cd3fcbd0fc03" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.31:8080/\": dial tcp 10.217.0.31:8080: connect: connection refused" Dec 03 16:05:15 crc kubenswrapper[4998]: E1203 16:05:15.232562 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:15.732544037 +0000 UTC m=+94.344244260 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.240065 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bbss6" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.243006 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wfm8s" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.247248 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-v8c75" podStartSLOduration=74.247232033 podStartE2EDuration="1m14.247232033s" podCreationTimestamp="2025-12-03 16:04:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:15.239264866 +0000 UTC m=+93.850965089" watchObservedRunningTime="2025-12-03 16:05:15.247232033 +0000 UTC m=+93.858932276" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.270080 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rrj4k" podStartSLOduration=75.270062558 podStartE2EDuration="1m15.270062558s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:15.255786212 +0000 UTC m=+93.867486435" watchObservedRunningTime="2025-12-03 16:05:15.270062558 +0000 UTC m=+93.881762771" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.271121 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-b9f5r" podStartSLOduration=7.271117452 podStartE2EDuration="7.271117452s" podCreationTimestamp="2025-12-03 16:05:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:15.269770572 +0000 UTC m=+93.881470795" watchObservedRunningTime="2025-12-03 16:05:15.271117452 +0000 UTC m=+93.882817675" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.289256 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-fnxt7" podStartSLOduration=74.289242773 podStartE2EDuration="1m14.289242773s" podCreationTimestamp="2025-12-03 16:04:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:15.287145057 +0000 UTC m=+93.898845280" watchObservedRunningTime="2025-12-03 16:05:15.289242773 +0000 UTC m=+93.900942996" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.302886 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-6vr8n" podStartSLOduration=75.302867425 podStartE2EDuration="1m15.302867425s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:15.301067855 +0000 UTC m=+93.912768088" watchObservedRunningTime="2025-12-03 16:05:15.302867425 +0000 UTC m=+93.914567648" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.318270 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5wc9c" podStartSLOduration=75.318255006 podStartE2EDuration="1m15.318255006s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:15.31617308 +0000 UTC m=+93.927873303" watchObservedRunningTime="2025-12-03 16:05:15.318255006 +0000 UTC m=+93.929955229" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.336636 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:15 crc kubenswrapper[4998]: E1203 16:05:15.342471 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:15.842455722 +0000 UTC m=+94.454155945 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.373702 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xd7m2" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.410274 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-gbvbf" podStartSLOduration=7.410254604 podStartE2EDuration="7.410254604s" podCreationTimestamp="2025-12-03 16:05:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:15.384075944 +0000 UTC m=+93.995776167" watchObservedRunningTime="2025-12-03 16:05:15.410254604 +0000 UTC m=+94.021954827" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.435276 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zh6tw" podStartSLOduration=74.435262298 podStartE2EDuration="1m14.435262298s" podCreationTimestamp="2025-12-03 16:04:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:15.412868642 +0000 UTC m=+94.024568865" watchObservedRunningTime="2025-12-03 16:05:15.435262298 +0000 UTC m=+94.046962521" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.439542 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:15 crc kubenswrapper[4998]: E1203 16:05:15.439942 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:15.939927831 +0000 UTC m=+94.551628054 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.475134 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-k92zl" podStartSLOduration=74.47511695 podStartE2EDuration="1m14.47511695s" podCreationTimestamp="2025-12-03 16:04:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:15.47509158 +0000 UTC m=+94.086791803" watchObservedRunningTime="2025-12-03 16:05:15.47511695 +0000 UTC m=+94.086817173" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.497722 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-5c2lp" podStartSLOduration=75.497705781 podStartE2EDuration="1m15.497705781s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:15.496804281 +0000 UTC m=+94.108504504" watchObservedRunningTime="2025-12-03 16:05:15.497705781 +0000 UTC m=+94.109406004" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.521360 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-jjwjb" podStartSLOduration=75.521345624 podStartE2EDuration="1m15.521345624s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:15.519717208 +0000 UTC m=+94.131417431" watchObservedRunningTime="2025-12-03 16:05:15.521345624 +0000 UTC m=+94.133045847" Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.541112 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:15 crc kubenswrapper[4998]: E1203 16:05:15.541480 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:16.04146429 +0000 UTC m=+94.653164513 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.642587 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:15 crc kubenswrapper[4998]: E1203 16:05:15.642977 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:16.142961618 +0000 UTC m=+94.754661841 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.744284 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:15 crc kubenswrapper[4998]: E1203 16:05:15.744635 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:16.24461776 +0000 UTC m=+94.856317983 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.845454 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:15 crc kubenswrapper[4998]: E1203 16:05:15.845767 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:16.34573858 +0000 UTC m=+94.957438803 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:15 crc kubenswrapper[4998]: I1203 16:05:15.947336 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:15 crc kubenswrapper[4998]: E1203 16:05:15.947644 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:16.447631697 +0000 UTC m=+95.059331920 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.048917 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:16 crc kubenswrapper[4998]: E1203 16:05:16.049139 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:16.549102114 +0000 UTC m=+95.160802337 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.049234 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:16 crc kubenswrapper[4998]: E1203 16:05:16.049559 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:16.549548334 +0000 UTC m=+95.161248557 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.141988 4998 patch_prober.go:28] interesting pod/router-default-5444994796-hnvvp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 16:05:16 crc kubenswrapper[4998]: [-]has-synced failed: reason withheld Dec 03 16:05:16 crc kubenswrapper[4998]: [+]process-running ok Dec 03 16:05:16 crc kubenswrapper[4998]: healthz check failed Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.142049 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hnvvp" podUID="cbae1035-ec59-4f24-884d-809c7e7bb685" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.150544 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:16 crc kubenswrapper[4998]: E1203 16:05:16.150747 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:16.650718005 +0000 UTC m=+95.262418228 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.235889 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jmdsz" event={"ID":"7274d4ec-5476-46c7-ab6a-f4ed0f4b294e","Type":"ContainerStarted","Data":"1e3b5a3cf6470a7a936dc3f6f052c70a281f82b734e321f9fe9f39811a9d3ff0"} Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.236275 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jmdsz" Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.237485 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-4rq7l" event={"ID":"57c54ccc-a846-4c00-8318-5d5331c70e70","Type":"ContainerStarted","Data":"4c45c01f121914209a6d0d1940f5b6e1deefaa17b6860b60875a5c5c7a1916d3"} Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.238533 4998 generic.go:334] "Generic (PLEG): container finished" podID="a6e6bbd8-c854-4708-b9b1-40884d1be6c7" containerID="481b1802323e8456ca60f05e1a85214ce7a71c0cba6ebdbe39be6ec00757d272" exitCode=0 Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.238577 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-6vr8n" event={"ID":"a6e6bbd8-c854-4708-b9b1-40884d1be6c7","Type":"ContainerDied","Data":"481b1802323e8456ca60f05e1a85214ce7a71c0cba6ebdbe39be6ec00757d272"} Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.239966 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-r2gbh" event={"ID":"22c37be4-5e07-4638-aa01-04ee06c463d7","Type":"ContainerStarted","Data":"a0295d1843c7bce89f475f09b713816f3cd008cc081cc26a8d15a797188b0bfe"} Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.241978 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dc55h" event={"ID":"f089be95-2900-436d-b1f4-627095e2d563","Type":"ContainerStarted","Data":"6c748016b0d34a10606e68f41522f362657c97969568c06bd8b35813e6a9ec17"} Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.244608 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5mthf" event={"ID":"e7a61bf3-1760-4910-9561-000d68ca9708","Type":"ContainerStarted","Data":"8eabbb1fe9d28b4977f96340aad7d554e6bc7365d2fa3245483661893b5d9f77"} Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.246992 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bhj2p" event={"ID":"ef0bf68e-f7d5-43a2-890c-10ffba3e6238","Type":"ContainerStarted","Data":"b7b4e50b1a0ea9c729f9c6618dc7aceb1a8eaf3f149a304bdfb54d2534a03183"} Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.265675 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jmdsz" podStartSLOduration=75.265660012 podStartE2EDuration="1m15.265660012s" podCreationTimestamp="2025-12-03 16:04:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:16.265283383 +0000 UTC m=+94.876983606" watchObservedRunningTime="2025-12-03 16:05:16.265660012 +0000 UTC m=+94.877360235" Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.266278 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.266598 4998 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-fnxt7 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.266647 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-fnxt7" podUID="8c715a96-664c-4b26-8d5a-523a0fb4a17c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" Dec 03 16:05:16 crc kubenswrapper[4998]: E1203 16:05:16.266658 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:16.766643813 +0000 UTC m=+95.378344036 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.322771 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dc55h" podStartSLOduration=75.322741206 podStartE2EDuration="1m15.322741206s" podCreationTimestamp="2025-12-03 16:04:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:16.319859842 +0000 UTC m=+94.931560065" watchObservedRunningTime="2025-12-03 16:05:16.322741206 +0000 UTC m=+94.934441429" Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.323452 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-r2gbh" podStartSLOduration=75.323447232 podStartE2EDuration="1m15.323447232s" podCreationTimestamp="2025-12-03 16:04:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:16.30125169 +0000 UTC m=+94.912951913" watchObservedRunningTime="2025-12-03 16:05:16.323447232 +0000 UTC m=+94.935147455" Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.367161 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.367696 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-bhj2p" podStartSLOduration=76.367679641 podStartE2EDuration="1m16.367679641s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:16.366499365 +0000 UTC m=+94.978199588" watchObservedRunningTime="2025-12-03 16:05:16.367679641 +0000 UTC m=+94.979379864" Dec 03 16:05:16 crc kubenswrapper[4998]: E1203 16:05:16.369682 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:16.869657005 +0000 UTC m=+95.481357298 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.401047 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-4rq7l" podStartSLOduration=76.40103318 podStartE2EDuration="1m16.40103318s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:16.399487686 +0000 UTC m=+95.011187909" watchObservedRunningTime="2025-12-03 16:05:16.40103318 +0000 UTC m=+95.012733403" Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.469194 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:16 crc kubenswrapper[4998]: E1203 16:05:16.469504 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:16.969492247 +0000 UTC m=+95.581192470 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.570480 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:16 crc kubenswrapper[4998]: E1203 16:05:16.570822 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:17.070807351 +0000 UTC m=+95.682507574 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.671722 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:16 crc kubenswrapper[4998]: E1203 16:05:16.672192 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:17.172172566 +0000 UTC m=+95.783872789 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.773416 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:16 crc kubenswrapper[4998]: E1203 16:05:16.773546 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:17.273527961 +0000 UTC m=+95.885228184 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.773793 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:16 crc kubenswrapper[4998]: E1203 16:05:16.774205 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:17.274197976 +0000 UTC m=+95.885898199 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.875003 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:16 crc kubenswrapper[4998]: E1203 16:05:16.875228 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:17.375202483 +0000 UTC m=+95.986902706 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.898001 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zh6tw" Dec 03 16:05:16 crc kubenswrapper[4998]: I1203 16:05:16.976252 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:16 crc kubenswrapper[4998]: E1203 16:05:16.976646 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:17.47662824 +0000 UTC m=+96.088328463 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.017830 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-84rhg"] Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.019295 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-84rhg" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.022346 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.036185 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-84rhg"] Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.077402 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:17 crc kubenswrapper[4998]: E1203 16:05:17.077568 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:17.577543025 +0000 UTC m=+96.189243248 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.078045 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea-utilities\") pod \"certified-operators-84rhg\" (UID: \"8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea\") " pod="openshift-marketplace/certified-operators-84rhg" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.078297 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52sp8\" (UniqueName: \"kubernetes.io/projected/8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea-kube-api-access-52sp8\") pod \"certified-operators-84rhg\" (UID: \"8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea\") " pod="openshift-marketplace/certified-operators-84rhg" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.078362 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea-catalog-content\") pod \"certified-operators-84rhg\" (UID: \"8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea\") " pod="openshift-marketplace/certified-operators-84rhg" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.078393 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:17 crc kubenswrapper[4998]: E1203 16:05:17.078744 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:17.578729382 +0000 UTC m=+96.190429605 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.147480 4998 patch_prober.go:28] interesting pod/router-default-5444994796-hnvvp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 16:05:17 crc kubenswrapper[4998]: [-]has-synced failed: reason withheld Dec 03 16:05:17 crc kubenswrapper[4998]: [+]process-running ok Dec 03 16:05:17 crc kubenswrapper[4998]: healthz check failed Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.147548 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hnvvp" podUID="cbae1035-ec59-4f24-884d-809c7e7bb685" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.179840 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:17 crc kubenswrapper[4998]: E1203 16:05:17.180006 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:17.679982704 +0000 UTC m=+96.291682937 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.180182 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea-utilities\") pod \"certified-operators-84rhg\" (UID: \"8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea\") " pod="openshift-marketplace/certified-operators-84rhg" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.180254 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52sp8\" (UniqueName: \"kubernetes.io/projected/8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea-kube-api-access-52sp8\") pod \"certified-operators-84rhg\" (UID: \"8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea\") " pod="openshift-marketplace/certified-operators-84rhg" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.180307 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea-catalog-content\") pod \"certified-operators-84rhg\" (UID: \"8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea\") " pod="openshift-marketplace/certified-operators-84rhg" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.180335 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.180599 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea-utilities\") pod \"certified-operators-84rhg\" (UID: \"8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea\") " pod="openshift-marketplace/certified-operators-84rhg" Dec 03 16:05:17 crc kubenswrapper[4998]: E1203 16:05:17.180718 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:17.680708371 +0000 UTC m=+96.292408614 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.180910 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea-catalog-content\") pod \"certified-operators-84rhg\" (UID: \"8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea\") " pod="openshift-marketplace/certified-operators-84rhg" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.201795 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52sp8\" (UniqueName: \"kubernetes.io/projected/8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea-kube-api-access-52sp8\") pod \"certified-operators-84rhg\" (UID: \"8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea\") " pod="openshift-marketplace/certified-operators-84rhg" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.219410 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rrdtb"] Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.220617 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rrdtb" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.222414 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.234101 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rrdtb"] Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.265298 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5mthf" event={"ID":"e7a61bf3-1760-4910-9561-000d68ca9708","Type":"ContainerStarted","Data":"eff706a356cd520613f4645037b4673ae3760044a5267f3be688b07de9d89f81"} Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.266483 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5mthf" event={"ID":"e7a61bf3-1760-4910-9561-000d68ca9708","Type":"ContainerStarted","Data":"2ba971646f30407898a2a6fc3aefead22638b27b56f64e7281c2f85ea67065e0"} Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.270142 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kg4lw" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.282103 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.282401 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ef027be-08d9-4731-bd35-34f7bb2fe43a-catalog-content\") pod \"community-operators-rrdtb\" (UID: \"5ef027be-08d9-4731-bd35-34f7bb2fe43a\") " pod="openshift-marketplace/community-operators-rrdtb" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.282462 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkhch\" (UniqueName: \"kubernetes.io/projected/5ef027be-08d9-4731-bd35-34f7bb2fe43a-kube-api-access-dkhch\") pod \"community-operators-rrdtb\" (UID: \"5ef027be-08d9-4731-bd35-34f7bb2fe43a\") " pod="openshift-marketplace/community-operators-rrdtb" Dec 03 16:05:17 crc kubenswrapper[4998]: E1203 16:05:17.282502 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:17.782468965 +0000 UTC m=+96.394169188 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.282560 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.282701 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ef027be-08d9-4731-bd35-34f7bb2fe43a-utilities\") pod \"community-operators-rrdtb\" (UID: \"5ef027be-08d9-4731-bd35-34f7bb2fe43a\") " pod="openshift-marketplace/community-operators-rrdtb" Dec 03 16:05:17 crc kubenswrapper[4998]: E1203 16:05:17.283151 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:17.783134609 +0000 UTC m=+96.394834832 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.345113 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-84rhg" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.386079 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.386411 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ef027be-08d9-4731-bd35-34f7bb2fe43a-catalog-content\") pod \"community-operators-rrdtb\" (UID: \"5ef027be-08d9-4731-bd35-34f7bb2fe43a\") " pod="openshift-marketplace/community-operators-rrdtb" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.386515 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkhch\" (UniqueName: \"kubernetes.io/projected/5ef027be-08d9-4731-bd35-34f7bb2fe43a-kube-api-access-dkhch\") pod \"community-operators-rrdtb\" (UID: \"5ef027be-08d9-4731-bd35-34f7bb2fe43a\") " pod="openshift-marketplace/community-operators-rrdtb" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.386660 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ef027be-08d9-4731-bd35-34f7bb2fe43a-utilities\") pod \"community-operators-rrdtb\" (UID: \"5ef027be-08d9-4731-bd35-34f7bb2fe43a\") " pod="openshift-marketplace/community-operators-rrdtb" Dec 03 16:05:17 crc kubenswrapper[4998]: E1203 16:05:17.386926 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:17.886910938 +0000 UTC m=+96.498611151 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.388444 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ef027be-08d9-4731-bd35-34f7bb2fe43a-catalog-content\") pod \"community-operators-rrdtb\" (UID: \"5ef027be-08d9-4731-bd35-34f7bb2fe43a\") " pod="openshift-marketplace/community-operators-rrdtb" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.389033 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ef027be-08d9-4731-bd35-34f7bb2fe43a-utilities\") pod \"community-operators-rrdtb\" (UID: \"5ef027be-08d9-4731-bd35-34f7bb2fe43a\") " pod="openshift-marketplace/community-operators-rrdtb" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.422363 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-czz6d"] Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.424338 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-czz6d" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.432688 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkhch\" (UniqueName: \"kubernetes.io/projected/5ef027be-08d9-4731-bd35-34f7bb2fe43a-kube-api-access-dkhch\") pod \"community-operators-rrdtb\" (UID: \"5ef027be-08d9-4731-bd35-34f7bb2fe43a\") " pod="openshift-marketplace/community-operators-rrdtb" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.435314 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-czz6d"] Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.489969 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.490044 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrvbv\" (UniqueName: \"kubernetes.io/projected/50083a46-5fd2-4241-8a20-67ae6cd5f6b6-kube-api-access-vrvbv\") pod \"certified-operators-czz6d\" (UID: \"50083a46-5fd2-4241-8a20-67ae6cd5f6b6\") " pod="openshift-marketplace/certified-operators-czz6d" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.490131 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50083a46-5fd2-4241-8a20-67ae6cd5f6b6-catalog-content\") pod \"certified-operators-czz6d\" (UID: \"50083a46-5fd2-4241-8a20-67ae6cd5f6b6\") " pod="openshift-marketplace/certified-operators-czz6d" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.490159 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50083a46-5fd2-4241-8a20-67ae6cd5f6b6-utilities\") pod \"certified-operators-czz6d\" (UID: \"50083a46-5fd2-4241-8a20-67ae6cd5f6b6\") " pod="openshift-marketplace/certified-operators-czz6d" Dec 03 16:05:17 crc kubenswrapper[4998]: E1203 16:05:17.490475 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:17.990455752 +0000 UTC m=+96.602155975 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.499475 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-6vr8n" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.533900 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rrdtb" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.593372 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a6e6bbd8-c854-4708-b9b1-40884d1be6c7-secret-volume\") pod \"a6e6bbd8-c854-4708-b9b1-40884d1be6c7\" (UID: \"a6e6bbd8-c854-4708-b9b1-40884d1be6c7\") " Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.593530 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.593565 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a6e6bbd8-c854-4708-b9b1-40884d1be6c7-config-volume\") pod \"a6e6bbd8-c854-4708-b9b1-40884d1be6c7\" (UID: \"a6e6bbd8-c854-4708-b9b1-40884d1be6c7\") " Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.593618 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbj97\" (UniqueName: \"kubernetes.io/projected/a6e6bbd8-c854-4708-b9b1-40884d1be6c7-kube-api-access-lbj97\") pod \"a6e6bbd8-c854-4708-b9b1-40884d1be6c7\" (UID: \"a6e6bbd8-c854-4708-b9b1-40884d1be6c7\") " Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.593787 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrvbv\" (UniqueName: \"kubernetes.io/projected/50083a46-5fd2-4241-8a20-67ae6cd5f6b6-kube-api-access-vrvbv\") pod \"certified-operators-czz6d\" (UID: \"50083a46-5fd2-4241-8a20-67ae6cd5f6b6\") " pod="openshift-marketplace/certified-operators-czz6d" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.593857 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50083a46-5fd2-4241-8a20-67ae6cd5f6b6-catalog-content\") pod \"certified-operators-czz6d\" (UID: \"50083a46-5fd2-4241-8a20-67ae6cd5f6b6\") " pod="openshift-marketplace/certified-operators-czz6d" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.593876 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50083a46-5fd2-4241-8a20-67ae6cd5f6b6-utilities\") pod \"certified-operators-czz6d\" (UID: \"50083a46-5fd2-4241-8a20-67ae6cd5f6b6\") " pod="openshift-marketplace/certified-operators-czz6d" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.594720 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50083a46-5fd2-4241-8a20-67ae6cd5f6b6-utilities\") pod \"certified-operators-czz6d\" (UID: \"50083a46-5fd2-4241-8a20-67ae6cd5f6b6\") " pod="openshift-marketplace/certified-operators-czz6d" Dec 03 16:05:17 crc kubenswrapper[4998]: E1203 16:05:17.595105 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:18.095086709 +0000 UTC m=+96.706786932 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.595170 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50083a46-5fd2-4241-8a20-67ae6cd5f6b6-catalog-content\") pod \"certified-operators-czz6d\" (UID: \"50083a46-5fd2-4241-8a20-67ae6cd5f6b6\") " pod="openshift-marketplace/certified-operators-czz6d" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.596142 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6e6bbd8-c854-4708-b9b1-40884d1be6c7-config-volume" (OuterVolumeSpecName: "config-volume") pod "a6e6bbd8-c854-4708-b9b1-40884d1be6c7" (UID: "a6e6bbd8-c854-4708-b9b1-40884d1be6c7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.604116 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6e6bbd8-c854-4708-b9b1-40884d1be6c7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a6e6bbd8-c854-4708-b9b1-40884d1be6c7" (UID: "a6e6bbd8-c854-4708-b9b1-40884d1be6c7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.606111 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6e6bbd8-c854-4708-b9b1-40884d1be6c7-kube-api-access-lbj97" (OuterVolumeSpecName: "kube-api-access-lbj97") pod "a6e6bbd8-c854-4708-b9b1-40884d1be6c7" (UID: "a6e6bbd8-c854-4708-b9b1-40884d1be6c7"). InnerVolumeSpecName "kube-api-access-lbj97". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.618488 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5v5j8"] Dec 03 16:05:17 crc kubenswrapper[4998]: E1203 16:05:17.618676 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6e6bbd8-c854-4708-b9b1-40884d1be6c7" containerName="collect-profiles" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.618687 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6e6bbd8-c854-4708-b9b1-40884d1be6c7" containerName="collect-profiles" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.618845 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6e6bbd8-c854-4708-b9b1-40884d1be6c7" containerName="collect-profiles" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.619507 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrvbv\" (UniqueName: \"kubernetes.io/projected/50083a46-5fd2-4241-8a20-67ae6cd5f6b6-kube-api-access-vrvbv\") pod \"certified-operators-czz6d\" (UID: \"50083a46-5fd2-4241-8a20-67ae6cd5f6b6\") " pod="openshift-marketplace/certified-operators-czz6d" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.619598 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5v5j8" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.625350 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5v5j8"] Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.661976 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-84rhg"] Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.695541 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97fb2ff7-4894-4858-99aa-1a54a3119d33-utilities\") pod \"community-operators-5v5j8\" (UID: \"97fb2ff7-4894-4858-99aa-1a54a3119d33\") " pod="openshift-marketplace/community-operators-5v5j8" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.695921 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97fb2ff7-4894-4858-99aa-1a54a3119d33-catalog-content\") pod \"community-operators-5v5j8\" (UID: \"97fb2ff7-4894-4858-99aa-1a54a3119d33\") " pod="openshift-marketplace/community-operators-5v5j8" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.695994 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.696031 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcgtc\" (UniqueName: \"kubernetes.io/projected/97fb2ff7-4894-4858-99aa-1a54a3119d33-kube-api-access-hcgtc\") pod \"community-operators-5v5j8\" (UID: \"97fb2ff7-4894-4858-99aa-1a54a3119d33\") " pod="openshift-marketplace/community-operators-5v5j8" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.696084 4998 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a6e6bbd8-c854-4708-b9b1-40884d1be6c7-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.696095 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbj97\" (UniqueName: \"kubernetes.io/projected/a6e6bbd8-c854-4708-b9b1-40884d1be6c7-kube-api-access-lbj97\") on node \"crc\" DevicePath \"\"" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.696114 4998 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a6e6bbd8-c854-4708-b9b1-40884d1be6c7-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 16:05:17 crc kubenswrapper[4998]: E1203 16:05:17.696415 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 16:05:18.196402164 +0000 UTC m=+96.808102387 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-55d4c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.710309 4998 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.757298 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-czz6d" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.798243 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.798468 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcgtc\" (UniqueName: \"kubernetes.io/projected/97fb2ff7-4894-4858-99aa-1a54a3119d33-kube-api-access-hcgtc\") pod \"community-operators-5v5j8\" (UID: \"97fb2ff7-4894-4858-99aa-1a54a3119d33\") " pod="openshift-marketplace/community-operators-5v5j8" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.798527 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97fb2ff7-4894-4858-99aa-1a54a3119d33-utilities\") pod \"community-operators-5v5j8\" (UID: \"97fb2ff7-4894-4858-99aa-1a54a3119d33\") " pod="openshift-marketplace/community-operators-5v5j8" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.798562 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97fb2ff7-4894-4858-99aa-1a54a3119d33-catalog-content\") pod \"community-operators-5v5j8\" (UID: \"97fb2ff7-4894-4858-99aa-1a54a3119d33\") " pod="openshift-marketplace/community-operators-5v5j8" Dec 03 16:05:17 crc kubenswrapper[4998]: E1203 16:05:17.798794 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 16:05:18.298767401 +0000 UTC m=+96.910467624 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.799008 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97fb2ff7-4894-4858-99aa-1a54a3119d33-catalog-content\") pod \"community-operators-5v5j8\" (UID: \"97fb2ff7-4894-4858-99aa-1a54a3119d33\") " pod="openshift-marketplace/community-operators-5v5j8" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.799444 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97fb2ff7-4894-4858-99aa-1a54a3119d33-utilities\") pod \"community-operators-5v5j8\" (UID: \"97fb2ff7-4894-4858-99aa-1a54a3119d33\") " pod="openshift-marketplace/community-operators-5v5j8" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.802010 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rrdtb"] Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.812358 4998 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-03T16:05:17.710334942Z","Handler":null,"Name":""} Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.815874 4998 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.815939 4998 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.817367 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcgtc\" (UniqueName: \"kubernetes.io/projected/97fb2ff7-4894-4858-99aa-1a54a3119d33-kube-api-access-hcgtc\") pod \"community-operators-5v5j8\" (UID: \"97fb2ff7-4894-4858-99aa-1a54a3119d33\") " pod="openshift-marketplace/community-operators-5v5j8" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.899716 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.903596 4998 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.903626 4998 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.932500 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-55d4c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.940096 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5v5j8" Dec 03 16:05:17 crc kubenswrapper[4998]: I1203 16:05:17.952286 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-czz6d"] Dec 03 16:05:17 crc kubenswrapper[4998]: W1203 16:05:17.961669 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50083a46_5fd2_4241_8a20_67ae6cd5f6b6.slice/crio-6fe6b091ba814eb11f5421dca011d5d93fa0bfeba4c41534c59e756245cb8913 WatchSource:0}: Error finding container 6fe6b091ba814eb11f5421dca011d5d93fa0bfeba4c41534c59e756245cb8913: Status 404 returned error can't find the container with id 6fe6b091ba814eb11f5421dca011d5d93fa0bfeba4c41534c59e756245cb8913 Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.001983 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.009042 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.018669 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.128422 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5v5j8"] Dec 03 16:05:18 crc kubenswrapper[4998]: W1203 16:05:18.136209 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97fb2ff7_4894_4858_99aa_1a54a3119d33.slice/crio-12cde948b68a3f0837ca6d4e5518e0920bf4e842cb029d1cf8c9e3459c022eb1 WatchSource:0}: Error finding container 12cde948b68a3f0837ca6d4e5518e0920bf4e842cb029d1cf8c9e3459c022eb1: Status 404 returned error can't find the container with id 12cde948b68a3f0837ca6d4e5518e0920bf4e842cb029d1cf8c9e3459c022eb1 Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.140342 4998 patch_prober.go:28] interesting pod/router-default-5444994796-hnvvp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 16:05:18 crc kubenswrapper[4998]: [-]has-synced failed: reason withheld Dec 03 16:05:18 crc kubenswrapper[4998]: [+]process-running ok Dec 03 16:05:18 crc kubenswrapper[4998]: healthz check failed Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.140398 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hnvvp" podUID="cbae1035-ec59-4f24-884d-809c7e7bb685" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.201538 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-55d4c"] Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.204184 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2914eb52-522e-4789-a988-b15875755144-metrics-certs\") pod \"network-metrics-daemon-k8ptd\" (UID: \"2914eb52-522e-4789-a988-b15875755144\") " pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.208871 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2914eb52-522e-4789-a988-b15875755144-metrics-certs\") pod \"network-metrics-daemon-k8ptd\" (UID: \"2914eb52-522e-4789-a988-b15875755144\") " pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.274433 4998 generic.go:334] "Generic (PLEG): container finished" podID="50083a46-5fd2-4241-8a20-67ae6cd5f6b6" containerID="d29809dcd5982fb5233b79073def5e2ea6290e274c54d747abec4bf91e31ed27" exitCode=0 Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.274524 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-czz6d" event={"ID":"50083a46-5fd2-4241-8a20-67ae6cd5f6b6","Type":"ContainerDied","Data":"d29809dcd5982fb5233b79073def5e2ea6290e274c54d747abec4bf91e31ed27"} Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.274553 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-czz6d" event={"ID":"50083a46-5fd2-4241-8a20-67ae6cd5f6b6","Type":"ContainerStarted","Data":"6fe6b091ba814eb11f5421dca011d5d93fa0bfeba4c41534c59e756245cb8913"} Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.276203 4998 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.277440 4998 generic.go:334] "Generic (PLEG): container finished" podID="8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea" containerID="af8a03c1a0e3f68bb595014c7e56371b1f9573d230c086e614c3f3d94404140d" exitCode=0 Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.277477 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84rhg" event={"ID":"8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea","Type":"ContainerDied","Data":"af8a03c1a0e3f68bb595014c7e56371b1f9573d230c086e614c3f3d94404140d"} Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.277676 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84rhg" event={"ID":"8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea","Type":"ContainerStarted","Data":"e17423948c3a9e9cc421a30ee51b7690d2edd560fea445f6ae4760ec1423a981"} Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.279573 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" event={"ID":"e6207e94-fd68-464e-9830-895d96a3437c","Type":"ContainerStarted","Data":"5b11427285b9f80ee0e5488edfadb6c4e519e42ce1398edc3f8c155c201988a3"} Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.283659 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-6vr8n" event={"ID":"a6e6bbd8-c854-4708-b9b1-40884d1be6c7","Type":"ContainerDied","Data":"a26607d85381891eef17c236245f6e92b5a22cf8e6587d7a928a037fedf2b362"} Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.283718 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a26607d85381891eef17c236245f6e92b5a22cf8e6587d7a928a037fedf2b362" Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.283676 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412960-6vr8n" Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.286152 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5v5j8" event={"ID":"97fb2ff7-4894-4858-99aa-1a54a3119d33","Type":"ContainerStarted","Data":"12cde948b68a3f0837ca6d4e5518e0920bf4e842cb029d1cf8c9e3459c022eb1"} Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.290466 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5mthf" event={"ID":"e7a61bf3-1760-4910-9561-000d68ca9708","Type":"ContainerStarted","Data":"5233576ed6526a66c943abc4441572fa838e425c57ff711cf36e2237ab9afc01"} Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.291046 4998 generic.go:334] "Generic (PLEG): container finished" podID="5ef027be-08d9-4731-bd35-34f7bb2fe43a" containerID="0b499c108a7efc0fae07e669d4d60cf49c27eb8e5f18280f27cf13f90a4f0f49" exitCode=0 Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.291148 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rrdtb" event={"ID":"5ef027be-08d9-4731-bd35-34f7bb2fe43a","Type":"ContainerDied","Data":"0b499c108a7efc0fae07e669d4d60cf49c27eb8e5f18280f27cf13f90a4f0f49"} Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.291172 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rrdtb" event={"ID":"5ef027be-08d9-4731-bd35-34f7bb2fe43a","Type":"ContainerStarted","Data":"19bbba31d6c74a4d6a8681fb879b07e78290a2dd9ff873ca370c05447adc0050"} Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.370248 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-5mthf" podStartSLOduration=10.370225768 podStartE2EDuration="10.370225768s" podCreationTimestamp="2025-12-03 16:05:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:18.367479757 +0000 UTC m=+96.979179980" watchObservedRunningTime="2025-12-03 16:05:18.370225768 +0000 UTC m=+96.981926001" Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.392776 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k8ptd" Dec 03 16:05:18 crc kubenswrapper[4998]: I1203 16:05:18.576263 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-k8ptd"] Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.143423 4998 patch_prober.go:28] interesting pod/router-default-5444994796-hnvvp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 16:05:19 crc kubenswrapper[4998]: [-]has-synced failed: reason withheld Dec 03 16:05:19 crc kubenswrapper[4998]: [+]process-running ok Dec 03 16:05:19 crc kubenswrapper[4998]: healthz check failed Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.143706 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hnvvp" podUID="cbae1035-ec59-4f24-884d-809c7e7bb685" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.320813 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-k8ptd" event={"ID":"2914eb52-522e-4789-a988-b15875755144","Type":"ContainerStarted","Data":"9e798136746f61927abcd3a41aad37ed97559e023454dd3612ad6da01df3251a"} Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.320870 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-k8ptd" event={"ID":"2914eb52-522e-4789-a988-b15875755144","Type":"ContainerStarted","Data":"db212e75d5ab162b371ecf15e5eab14b068270d8af491f971c1b3927766fde46"} Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.322896 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" event={"ID":"e6207e94-fd68-464e-9830-895d96a3437c","Type":"ContainerStarted","Data":"70535b37817c6a127f7e248412618455c6549628e70e6a2ff99e6080d32be1a0"} Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.323009 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.326300 4998 generic.go:334] "Generic (PLEG): container finished" podID="97fb2ff7-4894-4858-99aa-1a54a3119d33" containerID="2d44e443563c313d23cad9b53b4cad1fbdec714b71d6264fcae19bc4db5281b0" exitCode=0 Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.327468 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5v5j8" event={"ID":"97fb2ff7-4894-4858-99aa-1a54a3119d33","Type":"ContainerDied","Data":"2d44e443563c313d23cad9b53b4cad1fbdec714b71d6264fcae19bc4db5281b0"} Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.344028 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" podStartSLOduration=79.344002958 podStartE2EDuration="1m19.344002958s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:19.341973363 +0000 UTC m=+97.953673576" watchObservedRunningTime="2025-12-03 16:05:19.344002958 +0000 UTC m=+97.955703181" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.416114 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dsh8q"] Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.418425 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dsh8q" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.420296 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.420351 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.422262 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.422545 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dsh8q"] Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.428248 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.526692 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85eaaffb-675d-4283-9ef9-4765097f8a51-utilities\") pod \"redhat-marketplace-dsh8q\" (UID: \"85eaaffb-675d-4283-9ef9-4765097f8a51\") " pod="openshift-marketplace/redhat-marketplace-dsh8q" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.527236 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85eaaffb-675d-4283-9ef9-4765097f8a51-catalog-content\") pod \"redhat-marketplace-dsh8q\" (UID: \"85eaaffb-675d-4283-9ef9-4765097f8a51\") " pod="openshift-marketplace/redhat-marketplace-dsh8q" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.527301 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwf58\" (UniqueName: \"kubernetes.io/projected/85eaaffb-675d-4283-9ef9-4765097f8a51-kube-api-access-mwf58\") pod \"redhat-marketplace-dsh8q\" (UID: \"85eaaffb-675d-4283-9ef9-4765097f8a51\") " pod="openshift-marketplace/redhat-marketplace-dsh8q" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.628798 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85eaaffb-675d-4283-9ef9-4765097f8a51-catalog-content\") pod \"redhat-marketplace-dsh8q\" (UID: \"85eaaffb-675d-4283-9ef9-4765097f8a51\") " pod="openshift-marketplace/redhat-marketplace-dsh8q" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.628862 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwf58\" (UniqueName: \"kubernetes.io/projected/85eaaffb-675d-4283-9ef9-4765097f8a51-kube-api-access-mwf58\") pod \"redhat-marketplace-dsh8q\" (UID: \"85eaaffb-675d-4283-9ef9-4765097f8a51\") " pod="openshift-marketplace/redhat-marketplace-dsh8q" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.628932 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85eaaffb-675d-4283-9ef9-4765097f8a51-utilities\") pod \"redhat-marketplace-dsh8q\" (UID: \"85eaaffb-675d-4283-9ef9-4765097f8a51\") " pod="openshift-marketplace/redhat-marketplace-dsh8q" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.629484 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85eaaffb-675d-4283-9ef9-4765097f8a51-utilities\") pod \"redhat-marketplace-dsh8q\" (UID: \"85eaaffb-675d-4283-9ef9-4765097f8a51\") " pod="openshift-marketplace/redhat-marketplace-dsh8q" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.630398 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85eaaffb-675d-4283-9ef9-4765097f8a51-catalog-content\") pod \"redhat-marketplace-dsh8q\" (UID: \"85eaaffb-675d-4283-9ef9-4765097f8a51\") " pod="openshift-marketplace/redhat-marketplace-dsh8q" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.658739 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwf58\" (UniqueName: \"kubernetes.io/projected/85eaaffb-675d-4283-9ef9-4765097f8a51-kube-api-access-mwf58\") pod \"redhat-marketplace-dsh8q\" (UID: \"85eaaffb-675d-4283-9ef9-4765097f8a51\") " pod="openshift-marketplace/redhat-marketplace-dsh8q" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.692839 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.762542 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dsh8q" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.812434 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c7jnc"] Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.814101 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c7jnc" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.821146 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c7jnc"] Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.939516 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d66602dc-ef95-4eec-989f-8e42dc4bcd02-catalog-content\") pod \"redhat-marketplace-c7jnc\" (UID: \"d66602dc-ef95-4eec-989f-8e42dc4bcd02\") " pod="openshift-marketplace/redhat-marketplace-c7jnc" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.939892 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d66602dc-ef95-4eec-989f-8e42dc4bcd02-utilities\") pod \"redhat-marketplace-c7jnc\" (UID: \"d66602dc-ef95-4eec-989f-8e42dc4bcd02\") " pod="openshift-marketplace/redhat-marketplace-c7jnc" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.939960 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwgsn\" (UniqueName: \"kubernetes.io/projected/d66602dc-ef95-4eec-989f-8e42dc4bcd02-kube-api-access-kwgsn\") pod \"redhat-marketplace-c7jnc\" (UID: \"d66602dc-ef95-4eec-989f-8e42dc4bcd02\") " pod="openshift-marketplace/redhat-marketplace-c7jnc" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.958985 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.959925 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.964638 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.964923 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 03 16:05:19 crc kubenswrapper[4998]: I1203 16:05:19.968940 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.016935 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dsh8q"] Dec 03 16:05:20 crc kubenswrapper[4998]: W1203 16:05:20.038503 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85eaaffb_675d_4283_9ef9_4765097f8a51.slice/crio-75950a1ada30242cdd65fa71d3945ab82a42c81a3da99dcec7e844fbd569e98c WatchSource:0}: Error finding container 75950a1ada30242cdd65fa71d3945ab82a42c81a3da99dcec7e844fbd569e98c: Status 404 returned error can't find the container with id 75950a1ada30242cdd65fa71d3945ab82a42c81a3da99dcec7e844fbd569e98c Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.041573 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwgsn\" (UniqueName: \"kubernetes.io/projected/d66602dc-ef95-4eec-989f-8e42dc4bcd02-kube-api-access-kwgsn\") pod \"redhat-marketplace-c7jnc\" (UID: \"d66602dc-ef95-4eec-989f-8e42dc4bcd02\") " pod="openshift-marketplace/redhat-marketplace-c7jnc" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.041665 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/90394f27-9778-4a72-a09b-a709d8279ef8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"90394f27-9778-4a72-a09b-a709d8279ef8\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.041704 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d66602dc-ef95-4eec-989f-8e42dc4bcd02-catalog-content\") pod \"redhat-marketplace-c7jnc\" (UID: \"d66602dc-ef95-4eec-989f-8e42dc4bcd02\") " pod="openshift-marketplace/redhat-marketplace-c7jnc" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.041803 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/90394f27-9778-4a72-a09b-a709d8279ef8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"90394f27-9778-4a72-a09b-a709d8279ef8\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.042712 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d66602dc-ef95-4eec-989f-8e42dc4bcd02-catalog-content\") pod \"redhat-marketplace-c7jnc\" (UID: \"d66602dc-ef95-4eec-989f-8e42dc4bcd02\") " pod="openshift-marketplace/redhat-marketplace-c7jnc" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.042898 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d66602dc-ef95-4eec-989f-8e42dc4bcd02-utilities\") pod \"redhat-marketplace-c7jnc\" (UID: \"d66602dc-ef95-4eec-989f-8e42dc4bcd02\") " pod="openshift-marketplace/redhat-marketplace-c7jnc" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.043378 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d66602dc-ef95-4eec-989f-8e42dc4bcd02-utilities\") pod \"redhat-marketplace-c7jnc\" (UID: \"d66602dc-ef95-4eec-989f-8e42dc4bcd02\") " pod="openshift-marketplace/redhat-marketplace-c7jnc" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.077537 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwgsn\" (UniqueName: \"kubernetes.io/projected/d66602dc-ef95-4eec-989f-8e42dc4bcd02-kube-api-access-kwgsn\") pod \"redhat-marketplace-c7jnc\" (UID: \"d66602dc-ef95-4eec-989f-8e42dc4bcd02\") " pod="openshift-marketplace/redhat-marketplace-c7jnc" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.136641 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c7jnc" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.142474 4998 patch_prober.go:28] interesting pod/router-default-5444994796-hnvvp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 16:05:20 crc kubenswrapper[4998]: [-]has-synced failed: reason withheld Dec 03 16:05:20 crc kubenswrapper[4998]: [+]process-running ok Dec 03 16:05:20 crc kubenswrapper[4998]: healthz check failed Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.142532 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hnvvp" podUID="cbae1035-ec59-4f24-884d-809c7e7bb685" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.144162 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/90394f27-9778-4a72-a09b-a709d8279ef8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"90394f27-9778-4a72-a09b-a709d8279ef8\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.144545 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/90394f27-9778-4a72-a09b-a709d8279ef8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"90394f27-9778-4a72-a09b-a709d8279ef8\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.144620 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/90394f27-9778-4a72-a09b-a709d8279ef8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"90394f27-9778-4a72-a09b-a709d8279ef8\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.170699 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/90394f27-9778-4a72-a09b-a709d8279ef8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"90394f27-9778-4a72-a09b-a709d8279ef8\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.212976 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6rx7t"] Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.214283 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6rx7t" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.221647 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6rx7t"] Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.222540 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.289911 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.346619 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dsh8q" event={"ID":"85eaaffb-675d-4283-9ef9-4765097f8a51","Type":"ContainerStarted","Data":"75950a1ada30242cdd65fa71d3945ab82a42c81a3da99dcec7e844fbd569e98c"} Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.347099 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34-catalog-content\") pod \"redhat-operators-6rx7t\" (UID: \"ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34\") " pod="openshift-marketplace/redhat-operators-6rx7t" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.347124 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk5vk\" (UniqueName: \"kubernetes.io/projected/ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34-kube-api-access-xk5vk\") pod \"redhat-operators-6rx7t\" (UID: \"ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34\") " pod="openshift-marketplace/redhat-operators-6rx7t" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.347220 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34-utilities\") pod \"redhat-operators-6rx7t\" (UID: \"ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34\") " pod="openshift-marketplace/redhat-operators-6rx7t" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.353150 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-k8ptd" event={"ID":"2914eb52-522e-4789-a988-b15875755144","Type":"ContainerStarted","Data":"c554b58a7d6f3a3850a3f17bdd97fc269a2dc8e2718e5315ebdab5584ea012d8"} Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.357839 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-hwftp" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.368635 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-k8ptd" podStartSLOduration=80.368605574 podStartE2EDuration="1m20.368605574s" podCreationTimestamp="2025-12-03 16:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:20.367496299 +0000 UTC m=+98.979196522" watchObservedRunningTime="2025-12-03 16:05:20.368605574 +0000 UTC m=+98.980305797" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.429945 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-49jvv"] Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.441068 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-49jvv" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.456224 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-49jvv"] Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.462069 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34-utilities\") pod \"redhat-operators-6rx7t\" (UID: \"ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34\") " pod="openshift-marketplace/redhat-operators-6rx7t" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.462336 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34-catalog-content\") pod \"redhat-operators-6rx7t\" (UID: \"ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34\") " pod="openshift-marketplace/redhat-operators-6rx7t" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.462382 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk5vk\" (UniqueName: \"kubernetes.io/projected/ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34-kube-api-access-xk5vk\") pod \"redhat-operators-6rx7t\" (UID: \"ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34\") " pod="openshift-marketplace/redhat-operators-6rx7t" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.463173 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34-catalog-content\") pod \"redhat-operators-6rx7t\" (UID: \"ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34\") " pod="openshift-marketplace/redhat-operators-6rx7t" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.473067 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34-utilities\") pod \"redhat-operators-6rx7t\" (UID: \"ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34\") " pod="openshift-marketplace/redhat-operators-6rx7t" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.486914 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk5vk\" (UniqueName: \"kubernetes.io/projected/ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34-kube-api-access-xk5vk\") pod \"redhat-operators-6rx7t\" (UID: \"ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34\") " pod="openshift-marketplace/redhat-operators-6rx7t" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.553568 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6rx7t" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.567614 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69fee984-b61f-4a09-b330-c4aee6a2e41b-catalog-content\") pod \"redhat-operators-49jvv\" (UID: \"69fee984-b61f-4a09-b330-c4aee6a2e41b\") " pod="openshift-marketplace/redhat-operators-49jvv" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.567680 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69fee984-b61f-4a09-b330-c4aee6a2e41b-utilities\") pod \"redhat-operators-49jvv\" (UID: \"69fee984-b61f-4a09-b330-c4aee6a2e41b\") " pod="openshift-marketplace/redhat-operators-49jvv" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.567741 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vnbx\" (UniqueName: \"kubernetes.io/projected/69fee984-b61f-4a09-b330-c4aee6a2e41b-kube-api-access-8vnbx\") pod \"redhat-operators-49jvv\" (UID: \"69fee984-b61f-4a09-b330-c4aee6a2e41b\") " pod="openshift-marketplace/redhat-operators-49jvv" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.669550 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vnbx\" (UniqueName: \"kubernetes.io/projected/69fee984-b61f-4a09-b330-c4aee6a2e41b-kube-api-access-8vnbx\") pod \"redhat-operators-49jvv\" (UID: \"69fee984-b61f-4a09-b330-c4aee6a2e41b\") " pod="openshift-marketplace/redhat-operators-49jvv" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.669665 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69fee984-b61f-4a09-b330-c4aee6a2e41b-catalog-content\") pod \"redhat-operators-49jvv\" (UID: \"69fee984-b61f-4a09-b330-c4aee6a2e41b\") " pod="openshift-marketplace/redhat-operators-49jvv" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.669690 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69fee984-b61f-4a09-b330-c4aee6a2e41b-utilities\") pod \"redhat-operators-49jvv\" (UID: \"69fee984-b61f-4a09-b330-c4aee6a2e41b\") " pod="openshift-marketplace/redhat-operators-49jvv" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.670289 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69fee984-b61f-4a09-b330-c4aee6a2e41b-utilities\") pod \"redhat-operators-49jvv\" (UID: \"69fee984-b61f-4a09-b330-c4aee6a2e41b\") " pod="openshift-marketplace/redhat-operators-49jvv" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.670772 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69fee984-b61f-4a09-b330-c4aee6a2e41b-catalog-content\") pod \"redhat-operators-49jvv\" (UID: \"69fee984-b61f-4a09-b330-c4aee6a2e41b\") " pod="openshift-marketplace/redhat-operators-49jvv" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.687780 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vnbx\" (UniqueName: \"kubernetes.io/projected/69fee984-b61f-4a09-b330-c4aee6a2e41b-kube-api-access-8vnbx\") pod \"redhat-operators-49jvv\" (UID: \"69fee984-b61f-4a09-b330-c4aee6a2e41b\") " pod="openshift-marketplace/redhat-operators-49jvv" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.713659 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.714287 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.718220 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.718490 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.728603 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.728648 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.728662 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.743809 4998 patch_prober.go:28] interesting pod/console-f9d7485db-kk8mb container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.30:8443/health\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.743855 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-kk8mb" podUID="03021110-7732-4dff-a5e3-f481e8b7c0d6" containerName="console" probeResult="failure" output="Get \"https://10.217.0.30:8443/health\": dial tcp 10.217.0.30:8443: connect: connection refused" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.744353 4998 patch_prober.go:28] interesting pod/downloads-7954f5f757-ppw5v container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.744390 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ppw5v" podUID="2e58b61f-82bf-42c2-a664-cd3fcbd0fc03" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.31:8080/\": dial tcp 10.217.0.31:8080: connect: connection refused" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.744398 4998 patch_prober.go:28] interesting pod/downloads-7954f5f757-ppw5v container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.31:8080/\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.744441 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-ppw5v" podUID="2e58b61f-82bf-42c2-a664-cd3fcbd0fc03" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.31:8080/\": dial tcp 10.217.0.31:8080: connect: connection refused" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.745301 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c7jnc"] Dec 03 16:05:20 crc kubenswrapper[4998]: W1203 16:05:20.745816 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd66602dc_ef95_4eec_989f_8e42dc4bcd02.slice/crio-d6c4307c2f25ad7381a0a4105491285868587d480028b7d4d668ee9526a3fc6b WatchSource:0}: Error finding container d6c4307c2f25ad7381a0a4105491285868587d480028b7d4d668ee9526a3fc6b: Status 404 returned error can't find the container with id d6c4307c2f25ad7381a0a4105491285868587d480028b7d4d668ee9526a3fc6b Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.770884 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4628cf09-5693-4e77-8936-04ccd0f43a8b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4628cf09-5693-4e77-8936-04ccd0f43a8b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.770994 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4628cf09-5693-4e77-8936-04ccd0f43a8b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4628cf09-5693-4e77-8936-04ccd0f43a8b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.840075 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-49jvv" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.871817 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.872333 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4628cf09-5693-4e77-8936-04ccd0f43a8b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4628cf09-5693-4e77-8936-04ccd0f43a8b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.872415 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4628cf09-5693-4e77-8936-04ccd0f43a8b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4628cf09-5693-4e77-8936-04ccd0f43a8b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.875626 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4628cf09-5693-4e77-8936-04ccd0f43a8b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4628cf09-5693-4e77-8936-04ccd0f43a8b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 16:05:20 crc kubenswrapper[4998]: I1203 16:05:20.897081 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4628cf09-5693-4e77-8936-04ccd0f43a8b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4628cf09-5693-4e77-8936-04ccd0f43a8b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 16:05:21 crc kubenswrapper[4998]: I1203 16:05:21.037405 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6rx7t"] Dec 03 16:05:21 crc kubenswrapper[4998]: I1203 16:05:21.047112 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 16:05:21 crc kubenswrapper[4998]: W1203 16:05:21.107118 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea1d0431_ca9a_4275_8ef0_96e9ca7c0e34.slice/crio-bc8c7fa6598914723871453ecb9bb50e56ab5d0b3b5c8b45291b26715320066d WatchSource:0}: Error finding container bc8c7fa6598914723871453ecb9bb50e56ab5d0b3b5c8b45291b26715320066d: Status 404 returned error can't find the container with id bc8c7fa6598914723871453ecb9bb50e56ab5d0b3b5c8b45291b26715320066d Dec 03 16:05:21 crc kubenswrapper[4998]: I1203 16:05:21.138154 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-hnvvp" Dec 03 16:05:21 crc kubenswrapper[4998]: I1203 16:05:21.140688 4998 patch_prober.go:28] interesting pod/router-default-5444994796-hnvvp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 16:05:21 crc kubenswrapper[4998]: [-]has-synced failed: reason withheld Dec 03 16:05:21 crc kubenswrapper[4998]: [+]process-running ok Dec 03 16:05:21 crc kubenswrapper[4998]: healthz check failed Dec 03 16:05:21 crc kubenswrapper[4998]: I1203 16:05:21.140720 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hnvvp" podUID="cbae1035-ec59-4f24-884d-809c7e7bb685" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 16:05:21 crc kubenswrapper[4998]: I1203 16:05:21.174182 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-fnxt7" Dec 03 16:05:21 crc kubenswrapper[4998]: I1203 16:05:21.322659 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-49jvv"] Dec 03 16:05:21 crc kubenswrapper[4998]: I1203 16:05:21.356426 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 16:05:21 crc kubenswrapper[4998]: I1203 16:05:21.362959 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"90394f27-9778-4a72-a09b-a709d8279ef8","Type":"ContainerStarted","Data":"7fe1973b0e867d3a988c4ac6bdf8d106cb00d1f54fccbfd7648f03f900307d27"} Dec 03 16:05:21 crc kubenswrapper[4998]: I1203 16:05:21.364997 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6rx7t" event={"ID":"ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34","Type":"ContainerStarted","Data":"bc8c7fa6598914723871453ecb9bb50e56ab5d0b3b5c8b45291b26715320066d"} Dec 03 16:05:21 crc kubenswrapper[4998]: I1203 16:05:21.370063 4998 generic.go:334] "Generic (PLEG): container finished" podID="85eaaffb-675d-4283-9ef9-4765097f8a51" containerID="b13ec9ddb03a2d461bd4464b9402eed32b800bdd1e3d4fe4cebf0ec379f69ede" exitCode=0 Dec 03 16:05:21 crc kubenswrapper[4998]: I1203 16:05:21.370137 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dsh8q" event={"ID":"85eaaffb-675d-4283-9ef9-4765097f8a51","Type":"ContainerDied","Data":"b13ec9ddb03a2d461bd4464b9402eed32b800bdd1e3d4fe4cebf0ec379f69ede"} Dec 03 16:05:21 crc kubenswrapper[4998]: I1203 16:05:21.376038 4998 generic.go:334] "Generic (PLEG): container finished" podID="d66602dc-ef95-4eec-989f-8e42dc4bcd02" containerID="745579b66f1f14386edc835f36b411aad0b996a24fd1999a2116d23fa8a71646" exitCode=0 Dec 03 16:05:21 crc kubenswrapper[4998]: I1203 16:05:21.376842 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7jnc" event={"ID":"d66602dc-ef95-4eec-989f-8e42dc4bcd02","Type":"ContainerDied","Data":"745579b66f1f14386edc835f36b411aad0b996a24fd1999a2116d23fa8a71646"} Dec 03 16:05:21 crc kubenswrapper[4998]: I1203 16:05:21.376871 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7jnc" event={"ID":"d66602dc-ef95-4eec-989f-8e42dc4bcd02","Type":"ContainerStarted","Data":"d6c4307c2f25ad7381a0a4105491285868587d480028b7d4d668ee9526a3fc6b"} Dec 03 16:05:21 crc kubenswrapper[4998]: W1203 16:05:21.419458 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod4628cf09_5693_4e77_8936_04ccd0f43a8b.slice/crio-dec512974dbc4929ba50fbac82939e8dff8f5d4f134715a2a315d55c9c660a69 WatchSource:0}: Error finding container dec512974dbc4929ba50fbac82939e8dff8f5d4f134715a2a315d55c9c660a69: Status 404 returned error can't find the container with id dec512974dbc4929ba50fbac82939e8dff8f5d4f134715a2a315d55c9c660a69 Dec 03 16:05:22 crc kubenswrapper[4998]: I1203 16:05:22.136713 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:05:22 crc kubenswrapper[4998]: I1203 16:05:22.153054 4998 patch_prober.go:28] interesting pod/router-default-5444994796-hnvvp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 16:05:22 crc kubenswrapper[4998]: [-]has-synced failed: reason withheld Dec 03 16:05:22 crc kubenswrapper[4998]: [+]process-running ok Dec 03 16:05:22 crc kubenswrapper[4998]: healthz check failed Dec 03 16:05:22 crc kubenswrapper[4998]: I1203 16:05:22.153182 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-hnvvp" podUID="cbae1035-ec59-4f24-884d-809c7e7bb685" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 16:05:22 crc kubenswrapper[4998]: I1203 16:05:22.391554 4998 generic.go:334] "Generic (PLEG): container finished" podID="ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34" containerID="87f30a6956f51cf5e9b948c3d0533656461cabf031841ad2aaa6e49aa32d9dbc" exitCode=0 Dec 03 16:05:22 crc kubenswrapper[4998]: I1203 16:05:22.391659 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6rx7t" event={"ID":"ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34","Type":"ContainerDied","Data":"87f30a6956f51cf5e9b948c3d0533656461cabf031841ad2aaa6e49aa32d9dbc"} Dec 03 16:05:22 crc kubenswrapper[4998]: I1203 16:05:22.396632 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4628cf09-5693-4e77-8936-04ccd0f43a8b","Type":"ContainerStarted","Data":"a0a71fef780f830bd7dab3e5b9e0748762283cc32dc53d0fcf2fa0fb9caf03fd"} Dec 03 16:05:22 crc kubenswrapper[4998]: I1203 16:05:22.396675 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4628cf09-5693-4e77-8936-04ccd0f43a8b","Type":"ContainerStarted","Data":"dec512974dbc4929ba50fbac82939e8dff8f5d4f134715a2a315d55c9c660a69"} Dec 03 16:05:22 crc kubenswrapper[4998]: I1203 16:05:22.401787 4998 generic.go:334] "Generic (PLEG): container finished" podID="69fee984-b61f-4a09-b330-c4aee6a2e41b" containerID="98b086d9e385eab955edf149d812462ba6899cd9d24b7b00eb4900e435000798" exitCode=0 Dec 03 16:05:22 crc kubenswrapper[4998]: I1203 16:05:22.401875 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-49jvv" event={"ID":"69fee984-b61f-4a09-b330-c4aee6a2e41b","Type":"ContainerDied","Data":"98b086d9e385eab955edf149d812462ba6899cd9d24b7b00eb4900e435000798"} Dec 03 16:05:22 crc kubenswrapper[4998]: I1203 16:05:22.401927 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-49jvv" event={"ID":"69fee984-b61f-4a09-b330-c4aee6a2e41b","Type":"ContainerStarted","Data":"f094409fd5dc554e89d534706e20c2ca9eab0e24329d6c6a7b3eb31f074c5ead"} Dec 03 16:05:22 crc kubenswrapper[4998]: I1203 16:05:22.412022 4998 generic.go:334] "Generic (PLEG): container finished" podID="90394f27-9778-4a72-a09b-a709d8279ef8" containerID="cccbf9d24e66472ac0a15d2415bbfce8c276e47db517350dbd2c1845b6e1ca9f" exitCode=0 Dec 03 16:05:22 crc kubenswrapper[4998]: I1203 16:05:22.412239 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"90394f27-9778-4a72-a09b-a709d8279ef8","Type":"ContainerDied","Data":"cccbf9d24e66472ac0a15d2415bbfce8c276e47db517350dbd2c1845b6e1ca9f"} Dec 03 16:05:22 crc kubenswrapper[4998]: I1203 16:05:22.429344 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.429325169 podStartE2EDuration="2.429325169s" podCreationTimestamp="2025-12-03 16:05:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:05:22.427891877 +0000 UTC m=+101.039592100" watchObservedRunningTime="2025-12-03 16:05:22.429325169 +0000 UTC m=+101.041025392" Dec 03 16:05:23 crc kubenswrapper[4998]: I1203 16:05:23.140599 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-hnvvp" Dec 03 16:05:23 crc kubenswrapper[4998]: I1203 16:05:23.142880 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-hnvvp" Dec 03 16:05:23 crc kubenswrapper[4998]: I1203 16:05:23.686558 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 16:05:23 crc kubenswrapper[4998]: I1203 16:05:23.713478 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/90394f27-9778-4a72-a09b-a709d8279ef8-kube-api-access\") pod \"90394f27-9778-4a72-a09b-a709d8279ef8\" (UID: \"90394f27-9778-4a72-a09b-a709d8279ef8\") " Dec 03 16:05:23 crc kubenswrapper[4998]: I1203 16:05:23.713566 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/90394f27-9778-4a72-a09b-a709d8279ef8-kubelet-dir\") pod \"90394f27-9778-4a72-a09b-a709d8279ef8\" (UID: \"90394f27-9778-4a72-a09b-a709d8279ef8\") " Dec 03 16:05:23 crc kubenswrapper[4998]: I1203 16:05:23.713667 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/90394f27-9778-4a72-a09b-a709d8279ef8-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "90394f27-9778-4a72-a09b-a709d8279ef8" (UID: "90394f27-9778-4a72-a09b-a709d8279ef8"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:05:23 crc kubenswrapper[4998]: I1203 16:05:23.714433 4998 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/90394f27-9778-4a72-a09b-a709d8279ef8-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 16:05:23 crc kubenswrapper[4998]: I1203 16:05:23.719726 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90394f27-9778-4a72-a09b-a709d8279ef8-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "90394f27-9778-4a72-a09b-a709d8279ef8" (UID: "90394f27-9778-4a72-a09b-a709d8279ef8"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:05:23 crc kubenswrapper[4998]: I1203 16:05:23.815300 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/90394f27-9778-4a72-a09b-a709d8279ef8-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 16:05:24 crc kubenswrapper[4998]: I1203 16:05:24.431790 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"90394f27-9778-4a72-a09b-a709d8279ef8","Type":"ContainerDied","Data":"7fe1973b0e867d3a988c4ac6bdf8d106cb00d1f54fccbfd7648f03f900307d27"} Dec 03 16:05:24 crc kubenswrapper[4998]: I1203 16:05:24.432090 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7fe1973b0e867d3a988c4ac6bdf8d106cb00d1f54fccbfd7648f03f900307d27" Dec 03 16:05:24 crc kubenswrapper[4998]: I1203 16:05:24.431822 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 16:05:24 crc kubenswrapper[4998]: I1203 16:05:24.462379 4998 generic.go:334] "Generic (PLEG): container finished" podID="4628cf09-5693-4e77-8936-04ccd0f43a8b" containerID="a0a71fef780f830bd7dab3e5b9e0748762283cc32dc53d0fcf2fa0fb9caf03fd" exitCode=0 Dec 03 16:05:24 crc kubenswrapper[4998]: I1203 16:05:24.462417 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4628cf09-5693-4e77-8936-04ccd0f43a8b","Type":"ContainerDied","Data":"a0a71fef780f830bd7dab3e5b9e0748762283cc32dc53d0fcf2fa0fb9caf03fd"} Dec 03 16:05:25 crc kubenswrapper[4998]: I1203 16:05:25.995652 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-wsmkz" Dec 03 16:05:30 crc kubenswrapper[4998]: I1203 16:05:30.731143 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:30 crc kubenswrapper[4998]: I1203 16:05:30.737572 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:05:30 crc kubenswrapper[4998]: I1203 16:05:30.746546 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-ppw5v" Dec 03 16:05:31 crc kubenswrapper[4998]: I1203 16:05:31.545788 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 16:05:31 crc kubenswrapper[4998]: I1203 16:05:31.636700 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4628cf09-5693-4e77-8936-04ccd0f43a8b-kube-api-access\") pod \"4628cf09-5693-4e77-8936-04ccd0f43a8b\" (UID: \"4628cf09-5693-4e77-8936-04ccd0f43a8b\") " Dec 03 16:05:31 crc kubenswrapper[4998]: I1203 16:05:31.636813 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4628cf09-5693-4e77-8936-04ccd0f43a8b-kubelet-dir\") pod \"4628cf09-5693-4e77-8936-04ccd0f43a8b\" (UID: \"4628cf09-5693-4e77-8936-04ccd0f43a8b\") " Dec 03 16:05:31 crc kubenswrapper[4998]: I1203 16:05:31.637066 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4628cf09-5693-4e77-8936-04ccd0f43a8b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4628cf09-5693-4e77-8936-04ccd0f43a8b" (UID: "4628cf09-5693-4e77-8936-04ccd0f43a8b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:05:31 crc kubenswrapper[4998]: I1203 16:05:31.646230 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4628cf09-5693-4e77-8936-04ccd0f43a8b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4628cf09-5693-4e77-8936-04ccd0f43a8b" (UID: "4628cf09-5693-4e77-8936-04ccd0f43a8b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:05:31 crc kubenswrapper[4998]: I1203 16:05:31.738969 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4628cf09-5693-4e77-8936-04ccd0f43a8b-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 16:05:31 crc kubenswrapper[4998]: I1203 16:05:31.739447 4998 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4628cf09-5693-4e77-8936-04ccd0f43a8b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 16:05:32 crc kubenswrapper[4998]: I1203 16:05:32.521187 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4628cf09-5693-4e77-8936-04ccd0f43a8b","Type":"ContainerDied","Data":"dec512974dbc4929ba50fbac82939e8dff8f5d4f134715a2a315d55c9c660a69"} Dec 03 16:05:32 crc kubenswrapper[4998]: I1203 16:05:32.521246 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dec512974dbc4929ba50fbac82939e8dff8f5d4f134715a2a315d55c9c660a69" Dec 03 16:05:32 crc kubenswrapper[4998]: I1203 16:05:32.521218 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 16:05:38 crc kubenswrapper[4998]: I1203 16:05:38.024585 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:05:51 crc kubenswrapper[4998]: I1203 16:05:51.238772 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jmdsz" Dec 03 16:05:53 crc kubenswrapper[4998]: E1203 16:05:53.825148 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 03 16:05:53 crc kubenswrapper[4998]: E1203 16:05:53.825709 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dkhch,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-rrdtb_openshift-marketplace(5ef027be-08d9-4731-bd35-34f7bb2fe43a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 16:05:53 crc kubenswrapper[4998]: E1203 16:05:53.827074 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-rrdtb" podUID="5ef027be-08d9-4731-bd35-34f7bb2fe43a" Dec 03 16:05:56 crc kubenswrapper[4998]: E1203 16:05:56.194703 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-rrdtb" podUID="5ef027be-08d9-4731-bd35-34f7bb2fe43a" Dec 03 16:05:56 crc kubenswrapper[4998]: E1203 16:05:56.280863 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 03 16:05:56 crc kubenswrapper[4998]: E1203 16:05:56.281085 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-52sp8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-84rhg_openshift-marketplace(8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 16:05:56 crc kubenswrapper[4998]: E1203 16:05:56.281197 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 03 16:05:56 crc kubenswrapper[4998]: E1203 16:05:56.281348 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hcgtc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-5v5j8_openshift-marketplace(97fb2ff7-4894-4858-99aa-1a54a3119d33): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 16:05:56 crc kubenswrapper[4998]: E1203 16:05:56.282291 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-84rhg" podUID="8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea" Dec 03 16:05:56 crc kubenswrapper[4998]: E1203 16:05:56.283426 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-5v5j8" podUID="97fb2ff7-4894-4858-99aa-1a54a3119d33" Dec 03 16:05:56 crc kubenswrapper[4998]: I1203 16:05:56.917770 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 16:05:56 crc kubenswrapper[4998]: E1203 16:05:56.918256 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90394f27-9778-4a72-a09b-a709d8279ef8" containerName="pruner" Dec 03 16:05:56 crc kubenswrapper[4998]: I1203 16:05:56.918267 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="90394f27-9778-4a72-a09b-a709d8279ef8" containerName="pruner" Dec 03 16:05:56 crc kubenswrapper[4998]: E1203 16:05:56.918277 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4628cf09-5693-4e77-8936-04ccd0f43a8b" containerName="pruner" Dec 03 16:05:56 crc kubenswrapper[4998]: I1203 16:05:56.918283 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="4628cf09-5693-4e77-8936-04ccd0f43a8b" containerName="pruner" Dec 03 16:05:56 crc kubenswrapper[4998]: I1203 16:05:56.918379 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="90394f27-9778-4a72-a09b-a709d8279ef8" containerName="pruner" Dec 03 16:05:56 crc kubenswrapper[4998]: I1203 16:05:56.918389 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="4628cf09-5693-4e77-8936-04ccd0f43a8b" containerName="pruner" Dec 03 16:05:56 crc kubenswrapper[4998]: I1203 16:05:56.922000 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 16:05:56 crc kubenswrapper[4998]: I1203 16:05:56.923630 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 16:05:56 crc kubenswrapper[4998]: I1203 16:05:56.923985 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 16:05:56 crc kubenswrapper[4998]: I1203 16:05:56.925297 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 16:05:57 crc kubenswrapper[4998]: I1203 16:05:57.002840 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5e4733c6-a2dd-4beb-8240-6e57afff93cb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5e4733c6-a2dd-4beb-8240-6e57afff93cb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 16:05:57 crc kubenswrapper[4998]: I1203 16:05:57.003080 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5e4733c6-a2dd-4beb-8240-6e57afff93cb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5e4733c6-a2dd-4beb-8240-6e57afff93cb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 16:05:57 crc kubenswrapper[4998]: I1203 16:05:57.104594 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5e4733c6-a2dd-4beb-8240-6e57afff93cb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5e4733c6-a2dd-4beb-8240-6e57afff93cb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 16:05:57 crc kubenswrapper[4998]: I1203 16:05:57.104719 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5e4733c6-a2dd-4beb-8240-6e57afff93cb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5e4733c6-a2dd-4beb-8240-6e57afff93cb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 16:05:57 crc kubenswrapper[4998]: I1203 16:05:57.104898 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5e4733c6-a2dd-4beb-8240-6e57afff93cb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5e4733c6-a2dd-4beb-8240-6e57afff93cb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 16:05:57 crc kubenswrapper[4998]: I1203 16:05:57.139073 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5e4733c6-a2dd-4beb-8240-6e57afff93cb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5e4733c6-a2dd-4beb-8240-6e57afff93cb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 16:05:57 crc kubenswrapper[4998]: I1203 16:05:57.248204 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 16:05:59 crc kubenswrapper[4998]: E1203 16:05:59.858283 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-5v5j8" podUID="97fb2ff7-4894-4858-99aa-1a54a3119d33" Dec 03 16:05:59 crc kubenswrapper[4998]: E1203 16:05:59.858370 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-84rhg" podUID="8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea" Dec 03 16:05:59 crc kubenswrapper[4998]: E1203 16:05:59.877350 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \"/var/tmp/container_images_storage959508118/2\": happened during read: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 16:05:59 crc kubenswrapper[4998]: E1203 16:05:59.877676 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kwgsn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-c7jnc_openshift-marketplace(d66602dc-ef95-4eec-989f-8e42dc4bcd02): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \"/var/tmp/container_images_storage959508118/2\": happened during read: context canceled" logger="UnhandledError" Dec 03 16:05:59 crc kubenswrapper[4998]: E1203 16:05:59.878976 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \\\"/var/tmp/container_images_storage959508118/2\\\": happened during read: context canceled\"" pod="openshift-marketplace/redhat-marketplace-c7jnc" podUID="d66602dc-ef95-4eec-989f-8e42dc4bcd02" Dec 03 16:06:00 crc kubenswrapper[4998]: E1203 16:06:00.899549 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 16:06:00 crc kubenswrapper[4998]: E1203 16:06:00.900144 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mwf58,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-dsh8q_openshift-marketplace(85eaaffb-675d-4283-9ef9-4765097f8a51): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 16:06:00 crc kubenswrapper[4998]: E1203 16:06:00.901459 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-dsh8q" podUID="85eaaffb-675d-4283-9ef9-4765097f8a51" Dec 03 16:06:01 crc kubenswrapper[4998]: E1203 16:06:01.386868 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 03 16:06:01 crc kubenswrapper[4998]: E1203 16:06:01.387031 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vrvbv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-czz6d_openshift-marketplace(50083a46-5fd2-4241-8a20-67ae6cd5f6b6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 16:06:01 crc kubenswrapper[4998]: E1203 16:06:01.388191 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-czz6d" podUID="50083a46-5fd2-4241-8a20-67ae6cd5f6b6" Dec 03 16:06:01 crc kubenswrapper[4998]: I1203 16:06:01.512491 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 16:06:01 crc kubenswrapper[4998]: I1203 16:06:01.513513 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:06:01 crc kubenswrapper[4998]: I1203 16:06:01.523936 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 16:06:01 crc kubenswrapper[4998]: I1203 16:06:01.568297 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/20a499f0-e2b8-4204-9ff1-019ad9c78479-var-lock\") pod \"installer-9-crc\" (UID: \"20a499f0-e2b8-4204-9ff1-019ad9c78479\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:06:01 crc kubenswrapper[4998]: I1203 16:06:01.568366 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/20a499f0-e2b8-4204-9ff1-019ad9c78479-kube-api-access\") pod \"installer-9-crc\" (UID: \"20a499f0-e2b8-4204-9ff1-019ad9c78479\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:06:01 crc kubenswrapper[4998]: I1203 16:06:01.568390 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/20a499f0-e2b8-4204-9ff1-019ad9c78479-kubelet-dir\") pod \"installer-9-crc\" (UID: \"20a499f0-e2b8-4204-9ff1-019ad9c78479\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:06:01 crc kubenswrapper[4998]: I1203 16:06:01.670049 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/20a499f0-e2b8-4204-9ff1-019ad9c78479-var-lock\") pod \"installer-9-crc\" (UID: \"20a499f0-e2b8-4204-9ff1-019ad9c78479\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:06:01 crc kubenswrapper[4998]: I1203 16:06:01.670157 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/20a499f0-e2b8-4204-9ff1-019ad9c78479-var-lock\") pod \"installer-9-crc\" (UID: \"20a499f0-e2b8-4204-9ff1-019ad9c78479\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:06:01 crc kubenswrapper[4998]: I1203 16:06:01.670611 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/20a499f0-e2b8-4204-9ff1-019ad9c78479-kube-api-access\") pod \"installer-9-crc\" (UID: \"20a499f0-e2b8-4204-9ff1-019ad9c78479\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:06:01 crc kubenswrapper[4998]: I1203 16:06:01.670944 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/20a499f0-e2b8-4204-9ff1-019ad9c78479-kubelet-dir\") pod \"installer-9-crc\" (UID: \"20a499f0-e2b8-4204-9ff1-019ad9c78479\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:06:01 crc kubenswrapper[4998]: I1203 16:06:01.671069 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/20a499f0-e2b8-4204-9ff1-019ad9c78479-kubelet-dir\") pod \"installer-9-crc\" (UID: \"20a499f0-e2b8-4204-9ff1-019ad9c78479\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:06:01 crc kubenswrapper[4998]: I1203 16:06:01.691696 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/20a499f0-e2b8-4204-9ff1-019ad9c78479-kube-api-access\") pod \"installer-9-crc\" (UID: \"20a499f0-e2b8-4204-9ff1-019ad9c78479\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:06:01 crc kubenswrapper[4998]: I1203 16:06:01.841550 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:06:03 crc kubenswrapper[4998]: E1203 16:06:03.075053 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-czz6d" podUID="50083a46-5fd2-4241-8a20-67ae6cd5f6b6" Dec 03 16:06:03 crc kubenswrapper[4998]: E1203 16:06:03.075428 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-dsh8q" podUID="85eaaffb-675d-4283-9ef9-4765097f8a51" Dec 03 16:06:03 crc kubenswrapper[4998]: E1203 16:06:03.105864 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 03 16:06:03 crc kubenswrapper[4998]: E1203 16:06:03.106302 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xk5vk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-6rx7t_openshift-marketplace(ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 16:06:03 crc kubenswrapper[4998]: E1203 16:06:03.107442 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-6rx7t" podUID="ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34" Dec 03 16:06:03 crc kubenswrapper[4998]: I1203 16:06:03.478668 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 16:06:03 crc kubenswrapper[4998]: W1203 16:06:03.487520 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod20a499f0_e2b8_4204_9ff1_019ad9c78479.slice/crio-93e55e5129b09ebd864a5f2fda6959349d62a322faff567e32258d48c7ee16f2 WatchSource:0}: Error finding container 93e55e5129b09ebd864a5f2fda6959349d62a322faff567e32258d48c7ee16f2: Status 404 returned error can't find the container with id 93e55e5129b09ebd864a5f2fda6959349d62a322faff567e32258d48c7ee16f2 Dec 03 16:06:03 crc kubenswrapper[4998]: I1203 16:06:03.523610 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 16:06:03 crc kubenswrapper[4998]: W1203 16:06:03.532332 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod5e4733c6_a2dd_4beb_8240_6e57afff93cb.slice/crio-7702e40dc396b70cf5cefb7f39d0b186b09e610e16272ce7ddc426fa7ab53a8b WatchSource:0}: Error finding container 7702e40dc396b70cf5cefb7f39d0b186b09e610e16272ce7ddc426fa7ab53a8b: Status 404 returned error can't find the container with id 7702e40dc396b70cf5cefb7f39d0b186b09e610e16272ce7ddc426fa7ab53a8b Dec 03 16:06:03 crc kubenswrapper[4998]: I1203 16:06:03.712601 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"5e4733c6-a2dd-4beb-8240-6e57afff93cb","Type":"ContainerStarted","Data":"7702e40dc396b70cf5cefb7f39d0b186b09e610e16272ce7ddc426fa7ab53a8b"} Dec 03 16:06:03 crc kubenswrapper[4998]: I1203 16:06:03.714278 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-49jvv" event={"ID":"69fee984-b61f-4a09-b330-c4aee6a2e41b","Type":"ContainerStarted","Data":"14b61b88c5a5cfed8f79f27cf66ce7484e02b65e19279a0a305e28ca2fd90c3b"} Dec 03 16:06:03 crc kubenswrapper[4998]: I1203 16:06:03.715717 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"20a499f0-e2b8-4204-9ff1-019ad9c78479","Type":"ContainerStarted","Data":"93e55e5129b09ebd864a5f2fda6959349d62a322faff567e32258d48c7ee16f2"} Dec 03 16:06:03 crc kubenswrapper[4998]: E1203 16:06:03.718388 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-6rx7t" podUID="ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34" Dec 03 16:06:04 crc kubenswrapper[4998]: I1203 16:06:04.724285 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"20a499f0-e2b8-4204-9ff1-019ad9c78479","Type":"ContainerStarted","Data":"54c90d33a53ba3b54993ecc60273827e96597f341b9a19834c54af67e3a389da"} Dec 03 16:06:04 crc kubenswrapper[4998]: I1203 16:06:04.726395 4998 generic.go:334] "Generic (PLEG): container finished" podID="5e4733c6-a2dd-4beb-8240-6e57afff93cb" containerID="dc6d70d4bea7514a1cd597eac523623f831ea26e32a7b708abc6cd359278ee0d" exitCode=0 Dec 03 16:06:04 crc kubenswrapper[4998]: I1203 16:06:04.726483 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"5e4733c6-a2dd-4beb-8240-6e57afff93cb","Type":"ContainerDied","Data":"dc6d70d4bea7514a1cd597eac523623f831ea26e32a7b708abc6cd359278ee0d"} Dec 03 16:06:04 crc kubenswrapper[4998]: I1203 16:06:04.729385 4998 generic.go:334] "Generic (PLEG): container finished" podID="69fee984-b61f-4a09-b330-c4aee6a2e41b" containerID="14b61b88c5a5cfed8f79f27cf66ce7484e02b65e19279a0a305e28ca2fd90c3b" exitCode=0 Dec 03 16:06:04 crc kubenswrapper[4998]: I1203 16:06:04.729429 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-49jvv" event={"ID":"69fee984-b61f-4a09-b330-c4aee6a2e41b","Type":"ContainerDied","Data":"14b61b88c5a5cfed8f79f27cf66ce7484e02b65e19279a0a305e28ca2fd90c3b"} Dec 03 16:06:04 crc kubenswrapper[4998]: I1203 16:06:04.741040 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=3.74102347 podStartE2EDuration="3.74102347s" podCreationTimestamp="2025-12-03 16:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:06:04.739799269 +0000 UTC m=+143.351499532" watchObservedRunningTime="2025-12-03 16:06:04.74102347 +0000 UTC m=+143.352723723" Dec 03 16:06:05 crc kubenswrapper[4998]: I1203 16:06:05.736490 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-49jvv" event={"ID":"69fee984-b61f-4a09-b330-c4aee6a2e41b","Type":"ContainerStarted","Data":"82b26ab9f9f702629d2a4d98ff4f407ab6fa7b25c433e7fea0c4f8d3edb03d22"} Dec 03 16:06:05 crc kubenswrapper[4998]: I1203 16:06:05.759938 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-49jvv" podStartSLOduration=2.665280518 podStartE2EDuration="45.759913967s" podCreationTimestamp="2025-12-03 16:05:20 +0000 UTC" firstStartedPulling="2025-12-03 16:05:22.403985398 +0000 UTC m=+101.015685621" lastFinishedPulling="2025-12-03 16:06:05.498618807 +0000 UTC m=+144.110319070" observedRunningTime="2025-12-03 16:06:05.75766691 +0000 UTC m=+144.369367133" watchObservedRunningTime="2025-12-03 16:06:05.759913967 +0000 UTC m=+144.371614200" Dec 03 16:06:06 crc kubenswrapper[4998]: I1203 16:06:06.056034 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 16:06:06 crc kubenswrapper[4998]: I1203 16:06:06.135622 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5e4733c6-a2dd-4beb-8240-6e57afff93cb-kube-api-access\") pod \"5e4733c6-a2dd-4beb-8240-6e57afff93cb\" (UID: \"5e4733c6-a2dd-4beb-8240-6e57afff93cb\") " Dec 03 16:06:06 crc kubenswrapper[4998]: I1203 16:06:06.136050 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5e4733c6-a2dd-4beb-8240-6e57afff93cb-kubelet-dir\") pod \"5e4733c6-a2dd-4beb-8240-6e57afff93cb\" (UID: \"5e4733c6-a2dd-4beb-8240-6e57afff93cb\") " Dec 03 16:06:06 crc kubenswrapper[4998]: I1203 16:06:06.136126 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5e4733c6-a2dd-4beb-8240-6e57afff93cb-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "5e4733c6-a2dd-4beb-8240-6e57afff93cb" (UID: "5e4733c6-a2dd-4beb-8240-6e57afff93cb"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:06:06 crc kubenswrapper[4998]: I1203 16:06:06.136419 4998 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5e4733c6-a2dd-4beb-8240-6e57afff93cb-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:06 crc kubenswrapper[4998]: I1203 16:06:06.143968 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e4733c6-a2dd-4beb-8240-6e57afff93cb-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "5e4733c6-a2dd-4beb-8240-6e57afff93cb" (UID: "5e4733c6-a2dd-4beb-8240-6e57afff93cb"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:06:06 crc kubenswrapper[4998]: I1203 16:06:06.237497 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5e4733c6-a2dd-4beb-8240-6e57afff93cb-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:06 crc kubenswrapper[4998]: I1203 16:06:06.744764 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"5e4733c6-a2dd-4beb-8240-6e57afff93cb","Type":"ContainerDied","Data":"7702e40dc396b70cf5cefb7f39d0b186b09e610e16272ce7ddc426fa7ab53a8b"} Dec 03 16:06:06 crc kubenswrapper[4998]: I1203 16:06:06.744817 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7702e40dc396b70cf5cefb7f39d0b186b09e610e16272ce7ddc426fa7ab53a8b" Dec 03 16:06:06 crc kubenswrapper[4998]: I1203 16:06:06.744839 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 16:06:08 crc kubenswrapper[4998]: I1203 16:06:08.682070 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:06:08 crc kubenswrapper[4998]: I1203 16:06:08.683591 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:06:08 crc kubenswrapper[4998]: I1203 16:06:08.683956 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:06:08 crc kubenswrapper[4998]: I1203 16:06:08.684246 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:06:08 crc kubenswrapper[4998]: I1203 16:06:08.686344 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 16:06:08 crc kubenswrapper[4998]: I1203 16:06:08.687035 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 16:06:08 crc kubenswrapper[4998]: I1203 16:06:08.687181 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 16:06:08 crc kubenswrapper[4998]: I1203 16:06:08.696681 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:06:08 crc kubenswrapper[4998]: I1203 16:06:08.698071 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 16:06:08 crc kubenswrapper[4998]: I1203 16:06:08.702583 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:06:08 crc kubenswrapper[4998]: I1203 16:06:08.710672 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 16:06:08 crc kubenswrapper[4998]: I1203 16:06:08.716505 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:06:08 crc kubenswrapper[4998]: I1203 16:06:08.722077 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:06:08 crc kubenswrapper[4998]: I1203 16:06:08.990868 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 16:06:08 crc kubenswrapper[4998]: I1203 16:06:08.998716 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:06:09 crc kubenswrapper[4998]: W1203 16:06:09.530568 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-3da6c6050d21e3b2d29ffdb26ab31642a4987d364db5e221be155ed3e7f80c5e WatchSource:0}: Error finding container 3da6c6050d21e3b2d29ffdb26ab31642a4987d364db5e221be155ed3e7f80c5e: Status 404 returned error can't find the container with id 3da6c6050d21e3b2d29ffdb26ab31642a4987d364db5e221be155ed3e7f80c5e Dec 03 16:06:09 crc kubenswrapper[4998]: I1203 16:06:09.764117 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"47dae93e1858079d724bed114409a346d77e92c6df2af755acb6aac1a0ad9bc4"} Dec 03 16:06:09 crc kubenswrapper[4998]: I1203 16:06:09.764198 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"bf07719caca36f9dec6113a8d6b3ae3ef2827b8f8127f3dc882b222e496c0915"} Dec 03 16:06:09 crc kubenswrapper[4998]: I1203 16:06:09.764428 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:06:09 crc kubenswrapper[4998]: I1203 16:06:09.771855 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rrdtb" event={"ID":"5ef027be-08d9-4731-bd35-34f7bb2fe43a","Type":"ContainerStarted","Data":"17b15688ad76e3b67d4cdbe192aa2e3129305e78031fb82916e6295b9ad77ffa"} Dec 03 16:06:09 crc kubenswrapper[4998]: I1203 16:06:09.774306 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"a4ad18fd2b467f19964c159323d464003d5c6398b6028ece9cc3ef4d2530c846"} Dec 03 16:06:09 crc kubenswrapper[4998]: I1203 16:06:09.774359 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"3da6c6050d21e3b2d29ffdb26ab31642a4987d364db5e221be155ed3e7f80c5e"} Dec 03 16:06:09 crc kubenswrapper[4998]: I1203 16:06:09.775922 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"458e827c42f815ed5ba7d9a5c2f04e95e4fff936aa4fdb67aa42b005fcfa3141"} Dec 03 16:06:09 crc kubenswrapper[4998]: I1203 16:06:09.775958 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"d3782bcab8486ddfdd25e10535871ba223ccfe7aeccb388a45a00f4884363140"} Dec 03 16:06:10 crc kubenswrapper[4998]: I1203 16:06:10.783827 4998 generic.go:334] "Generic (PLEG): container finished" podID="5ef027be-08d9-4731-bd35-34f7bb2fe43a" containerID="17b15688ad76e3b67d4cdbe192aa2e3129305e78031fb82916e6295b9ad77ffa" exitCode=0 Dec 03 16:06:10 crc kubenswrapper[4998]: I1203 16:06:10.783907 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rrdtb" event={"ID":"5ef027be-08d9-4731-bd35-34f7bb2fe43a","Type":"ContainerDied","Data":"17b15688ad76e3b67d4cdbe192aa2e3129305e78031fb82916e6295b9ad77ffa"} Dec 03 16:06:10 crc kubenswrapper[4998]: I1203 16:06:10.841586 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-49jvv" Dec 03 16:06:10 crc kubenswrapper[4998]: I1203 16:06:10.841638 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-49jvv" Dec 03 16:06:11 crc kubenswrapper[4998]: I1203 16:06:11.793678 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rrdtb" event={"ID":"5ef027be-08d9-4731-bd35-34f7bb2fe43a","Type":"ContainerStarted","Data":"b4d411c20e1346b15e05f170e9d66fd9f1a449db05f1b05820f9b8dc0834578c"} Dec 03 16:06:11 crc kubenswrapper[4998]: I1203 16:06:11.817815 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rrdtb" podStartSLOduration=1.5993417810000001 podStartE2EDuration="54.81779694s" podCreationTimestamp="2025-12-03 16:05:17 +0000 UTC" firstStartedPulling="2025-12-03 16:05:18.293869217 +0000 UTC m=+96.905569440" lastFinishedPulling="2025-12-03 16:06:11.512324376 +0000 UTC m=+150.124024599" observedRunningTime="2025-12-03 16:06:11.814585989 +0000 UTC m=+150.426286252" watchObservedRunningTime="2025-12-03 16:06:11.81779694 +0000 UTC m=+150.429497163" Dec 03 16:06:11 crc kubenswrapper[4998]: I1203 16:06:11.906479 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-49jvv" podUID="69fee984-b61f-4a09-b330-c4aee6a2e41b" containerName="registry-server" probeResult="failure" output=< Dec 03 16:06:11 crc kubenswrapper[4998]: timeout: failed to connect service ":50051" within 1s Dec 03 16:06:11 crc kubenswrapper[4998]: > Dec 03 16:06:13 crc kubenswrapper[4998]: I1203 16:06:13.655535 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-b5hpf"] Dec 03 16:06:14 crc kubenswrapper[4998]: I1203 16:06:14.815039 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7jnc" event={"ID":"d66602dc-ef95-4eec-989f-8e42dc4bcd02","Type":"ContainerStarted","Data":"7e34c2ae31325ba2b7df3c95a9afd1a7e806e78ad412fa849dd99df19c6155b5"} Dec 03 16:06:15 crc kubenswrapper[4998]: I1203 16:06:15.822069 4998 generic.go:334] "Generic (PLEG): container finished" podID="d66602dc-ef95-4eec-989f-8e42dc4bcd02" containerID="7e34c2ae31325ba2b7df3c95a9afd1a7e806e78ad412fa849dd99df19c6155b5" exitCode=0 Dec 03 16:06:15 crc kubenswrapper[4998]: I1203 16:06:15.822218 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7jnc" event={"ID":"d66602dc-ef95-4eec-989f-8e42dc4bcd02","Type":"ContainerDied","Data":"7e34c2ae31325ba2b7df3c95a9afd1a7e806e78ad412fa849dd99df19c6155b5"} Dec 03 16:06:17 crc kubenswrapper[4998]: I1203 16:06:17.534920 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rrdtb" Dec 03 16:06:17 crc kubenswrapper[4998]: I1203 16:06:17.535697 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rrdtb" Dec 03 16:06:17 crc kubenswrapper[4998]: I1203 16:06:17.586679 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rrdtb" Dec 03 16:06:17 crc kubenswrapper[4998]: I1203 16:06:17.885249 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rrdtb" Dec 03 16:06:20 crc kubenswrapper[4998]: I1203 16:06:20.858137 4998 generic.go:334] "Generic (PLEG): container finished" podID="97fb2ff7-4894-4858-99aa-1a54a3119d33" containerID="5a2df0bd563fc1e5507c165e109d53270e713b31b63e9f1388677886d0e20d86" exitCode=0 Dec 03 16:06:20 crc kubenswrapper[4998]: I1203 16:06:20.858213 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5v5j8" event={"ID":"97fb2ff7-4894-4858-99aa-1a54a3119d33","Type":"ContainerDied","Data":"5a2df0bd563fc1e5507c165e109d53270e713b31b63e9f1388677886d0e20d86"} Dec 03 16:06:20 crc kubenswrapper[4998]: I1203 16:06:20.861766 4998 generic.go:334] "Generic (PLEG): container finished" podID="50083a46-5fd2-4241-8a20-67ae6cd5f6b6" containerID="d96a0239f749b70416e1f84fb2d6064e6d9c4413642bb00009e4984e08d80da5" exitCode=0 Dec 03 16:06:20 crc kubenswrapper[4998]: I1203 16:06:20.861799 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-czz6d" event={"ID":"50083a46-5fd2-4241-8a20-67ae6cd5f6b6","Type":"ContainerDied","Data":"d96a0239f749b70416e1f84fb2d6064e6d9c4413642bb00009e4984e08d80da5"} Dec 03 16:06:20 crc kubenswrapper[4998]: I1203 16:06:20.864970 4998 generic.go:334] "Generic (PLEG): container finished" podID="8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea" containerID="1d1830b0a1fe5b88ad88f60a31050e58f9f5b11241b7708d835b4462eedae371" exitCode=0 Dec 03 16:06:20 crc kubenswrapper[4998]: I1203 16:06:20.865039 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84rhg" event={"ID":"8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea","Type":"ContainerDied","Data":"1d1830b0a1fe5b88ad88f60a31050e58f9f5b11241b7708d835b4462eedae371"} Dec 03 16:06:20 crc kubenswrapper[4998]: I1203 16:06:20.869673 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6rx7t" event={"ID":"ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34","Type":"ContainerStarted","Data":"0c936724e87a152b1a722995b064278def6da26f462f12385824f59fef1feacf"} Dec 03 16:06:20 crc kubenswrapper[4998]: I1203 16:06:20.873701 4998 generic.go:334] "Generic (PLEG): container finished" podID="85eaaffb-675d-4283-9ef9-4765097f8a51" containerID="87c4e338a88366f2254f40595c16d184e8d890b446fb058042ca84b832367e78" exitCode=0 Dec 03 16:06:20 crc kubenswrapper[4998]: I1203 16:06:20.873788 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dsh8q" event={"ID":"85eaaffb-675d-4283-9ef9-4765097f8a51","Type":"ContainerDied","Data":"87c4e338a88366f2254f40595c16d184e8d890b446fb058042ca84b832367e78"} Dec 03 16:06:20 crc kubenswrapper[4998]: I1203 16:06:20.883969 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7jnc" event={"ID":"d66602dc-ef95-4eec-989f-8e42dc4bcd02","Type":"ContainerStarted","Data":"a9a5ccb5aff999ee23d08c44e58d8f36ce41d467a1ba4d5163add64d310f1c95"} Dec 03 16:06:20 crc kubenswrapper[4998]: I1203 16:06:20.902075 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-49jvv" Dec 03 16:06:20 crc kubenswrapper[4998]: I1203 16:06:20.952157 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-49jvv" Dec 03 16:06:20 crc kubenswrapper[4998]: I1203 16:06:20.975082 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c7jnc" podStartSLOduration=4.049332818 podStartE2EDuration="1m1.975062221s" podCreationTimestamp="2025-12-03 16:05:19 +0000 UTC" firstStartedPulling="2025-12-03 16:05:22.41896342 +0000 UTC m=+101.030663633" lastFinishedPulling="2025-12-03 16:06:20.344692813 +0000 UTC m=+158.956393036" observedRunningTime="2025-12-03 16:06:20.971395419 +0000 UTC m=+159.583095642" watchObservedRunningTime="2025-12-03 16:06:20.975062221 +0000 UTC m=+159.586762444" Dec 03 16:06:21 crc kubenswrapper[4998]: I1203 16:06:21.890082 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dsh8q" event={"ID":"85eaaffb-675d-4283-9ef9-4765097f8a51","Type":"ContainerStarted","Data":"d3a607b5ca6a8a42d2ef8df7393dc1263c11c7009612744e258b02823a97806f"} Dec 03 16:06:21 crc kubenswrapper[4998]: I1203 16:06:21.891879 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5v5j8" event={"ID":"97fb2ff7-4894-4858-99aa-1a54a3119d33","Type":"ContainerStarted","Data":"570e5c73768e5818016141650b8627643e932b16aa0ef1e5bad262256a9f109d"} Dec 03 16:06:21 crc kubenswrapper[4998]: I1203 16:06:21.894350 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-czz6d" event={"ID":"50083a46-5fd2-4241-8a20-67ae6cd5f6b6","Type":"ContainerStarted","Data":"1d31a6843d11f0c77a603504c4bfc4429fda271c8c4fbc39a9e25a04a012d09c"} Dec 03 16:06:21 crc kubenswrapper[4998]: I1203 16:06:21.896286 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84rhg" event={"ID":"8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea","Type":"ContainerStarted","Data":"08bd17c836a424fd644b4fbfe560085cf2d06aa7c9151a7a2bf0e1d29a30d34a"} Dec 03 16:06:21 crc kubenswrapper[4998]: I1203 16:06:21.898811 4998 generic.go:334] "Generic (PLEG): container finished" podID="ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34" containerID="0c936724e87a152b1a722995b064278def6da26f462f12385824f59fef1feacf" exitCode=0 Dec 03 16:06:21 crc kubenswrapper[4998]: I1203 16:06:21.898910 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6rx7t" event={"ID":"ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34","Type":"ContainerDied","Data":"0c936724e87a152b1a722995b064278def6da26f462f12385824f59fef1feacf"} Dec 03 16:06:21 crc kubenswrapper[4998]: I1203 16:06:21.898986 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6rx7t" event={"ID":"ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34","Type":"ContainerStarted","Data":"66b64f36635097dde5bf4df271ee7d55124f89c8c5c229fef9a8b1e3312645e8"} Dec 03 16:06:21 crc kubenswrapper[4998]: I1203 16:06:21.939647 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dsh8q" podStartSLOduration=2.711295753 podStartE2EDuration="1m2.939626268s" podCreationTimestamp="2025-12-03 16:05:19 +0000 UTC" firstStartedPulling="2025-12-03 16:05:21.412495227 +0000 UTC m=+100.024195450" lastFinishedPulling="2025-12-03 16:06:21.640825742 +0000 UTC m=+160.252525965" observedRunningTime="2025-12-03 16:06:21.913256753 +0000 UTC m=+160.524956986" watchObservedRunningTime="2025-12-03 16:06:21.939626268 +0000 UTC m=+160.551326491" Dec 03 16:06:21 crc kubenswrapper[4998]: I1203 16:06:21.942235 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5v5j8" podStartSLOduration=2.757463684 podStartE2EDuration="1m4.942227314s" podCreationTimestamp="2025-12-03 16:05:17 +0000 UTC" firstStartedPulling="2025-12-03 16:05:19.331764957 +0000 UTC m=+97.943465180" lastFinishedPulling="2025-12-03 16:06:21.516528587 +0000 UTC m=+160.128228810" observedRunningTime="2025-12-03 16:06:21.937523075 +0000 UTC m=+160.549223298" watchObservedRunningTime="2025-12-03 16:06:21.942227314 +0000 UTC m=+160.553927537" Dec 03 16:06:21 crc kubenswrapper[4998]: I1203 16:06:21.963190 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-czz6d" podStartSLOduration=1.609448441 podStartE2EDuration="1m4.963170702s" podCreationTimestamp="2025-12-03 16:05:17 +0000 UTC" firstStartedPulling="2025-12-03 16:05:18.275923059 +0000 UTC m=+96.887623282" lastFinishedPulling="2025-12-03 16:06:21.62964532 +0000 UTC m=+160.241345543" observedRunningTime="2025-12-03 16:06:21.961855709 +0000 UTC m=+160.573555932" watchObservedRunningTime="2025-12-03 16:06:21.963170702 +0000 UTC m=+160.574870925" Dec 03 16:06:22 crc kubenswrapper[4998]: I1203 16:06:22.015854 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-84rhg" podStartSLOduration=1.738218579 podStartE2EDuration="1m5.01583412s" podCreationTimestamp="2025-12-03 16:05:17 +0000 UTC" firstStartedPulling="2025-12-03 16:05:18.278817223 +0000 UTC m=+96.890517446" lastFinishedPulling="2025-12-03 16:06:21.556432764 +0000 UTC m=+160.168132987" observedRunningTime="2025-12-03 16:06:22.012534297 +0000 UTC m=+160.624234520" watchObservedRunningTime="2025-12-03 16:06:22.01583412 +0000 UTC m=+160.627534333" Dec 03 16:06:22 crc kubenswrapper[4998]: I1203 16:06:22.016891 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6rx7t" podStartSLOduration=2.686665522 podStartE2EDuration="1m2.016882327s" podCreationTimestamp="2025-12-03 16:05:20 +0000 UTC" firstStartedPulling="2025-12-03 16:05:22.396295278 +0000 UTC m=+101.007995501" lastFinishedPulling="2025-12-03 16:06:21.726512083 +0000 UTC m=+160.338212306" observedRunningTime="2025-12-03 16:06:21.989165838 +0000 UTC m=+160.600866071" watchObservedRunningTime="2025-12-03 16:06:22.016882327 +0000 UTC m=+160.628582580" Dec 03 16:06:22 crc kubenswrapper[4998]: I1203 16:06:22.819346 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-49jvv"] Dec 03 16:06:22 crc kubenswrapper[4998]: I1203 16:06:22.904099 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-49jvv" podUID="69fee984-b61f-4a09-b330-c4aee6a2e41b" containerName="registry-server" containerID="cri-o://82b26ab9f9f702629d2a4d98ff4f407ab6fa7b25c433e7fea0c4f8d3edb03d22" gracePeriod=2 Dec 03 16:06:23 crc kubenswrapper[4998]: I1203 16:06:23.300952 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-49jvv" Dec 03 16:06:23 crc kubenswrapper[4998]: I1203 16:06:23.390537 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69fee984-b61f-4a09-b330-c4aee6a2e41b-catalog-content\") pod \"69fee984-b61f-4a09-b330-c4aee6a2e41b\" (UID: \"69fee984-b61f-4a09-b330-c4aee6a2e41b\") " Dec 03 16:06:23 crc kubenswrapper[4998]: I1203 16:06:23.390589 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vnbx\" (UniqueName: \"kubernetes.io/projected/69fee984-b61f-4a09-b330-c4aee6a2e41b-kube-api-access-8vnbx\") pod \"69fee984-b61f-4a09-b330-c4aee6a2e41b\" (UID: \"69fee984-b61f-4a09-b330-c4aee6a2e41b\") " Dec 03 16:06:23 crc kubenswrapper[4998]: I1203 16:06:23.390615 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69fee984-b61f-4a09-b330-c4aee6a2e41b-utilities\") pod \"69fee984-b61f-4a09-b330-c4aee6a2e41b\" (UID: \"69fee984-b61f-4a09-b330-c4aee6a2e41b\") " Dec 03 16:06:23 crc kubenswrapper[4998]: I1203 16:06:23.391427 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69fee984-b61f-4a09-b330-c4aee6a2e41b-utilities" (OuterVolumeSpecName: "utilities") pod "69fee984-b61f-4a09-b330-c4aee6a2e41b" (UID: "69fee984-b61f-4a09-b330-c4aee6a2e41b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:06:23 crc kubenswrapper[4998]: I1203 16:06:23.396986 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69fee984-b61f-4a09-b330-c4aee6a2e41b-kube-api-access-8vnbx" (OuterVolumeSpecName: "kube-api-access-8vnbx") pod "69fee984-b61f-4a09-b330-c4aee6a2e41b" (UID: "69fee984-b61f-4a09-b330-c4aee6a2e41b"). InnerVolumeSpecName "kube-api-access-8vnbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:06:23 crc kubenswrapper[4998]: I1203 16:06:23.491932 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vnbx\" (UniqueName: \"kubernetes.io/projected/69fee984-b61f-4a09-b330-c4aee6a2e41b-kube-api-access-8vnbx\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:23 crc kubenswrapper[4998]: I1203 16:06:23.491963 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/69fee984-b61f-4a09-b330-c4aee6a2e41b-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:23 crc kubenswrapper[4998]: I1203 16:06:23.524239 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69fee984-b61f-4a09-b330-c4aee6a2e41b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "69fee984-b61f-4a09-b330-c4aee6a2e41b" (UID: "69fee984-b61f-4a09-b330-c4aee6a2e41b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:06:23 crc kubenswrapper[4998]: I1203 16:06:23.593715 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/69fee984-b61f-4a09-b330-c4aee6a2e41b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:23 crc kubenswrapper[4998]: I1203 16:06:23.917136 4998 generic.go:334] "Generic (PLEG): container finished" podID="69fee984-b61f-4a09-b330-c4aee6a2e41b" containerID="82b26ab9f9f702629d2a4d98ff4f407ab6fa7b25c433e7fea0c4f8d3edb03d22" exitCode=0 Dec 03 16:06:23 crc kubenswrapper[4998]: I1203 16:06:23.917208 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-49jvv" Dec 03 16:06:23 crc kubenswrapper[4998]: I1203 16:06:23.917245 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-49jvv" event={"ID":"69fee984-b61f-4a09-b330-c4aee6a2e41b","Type":"ContainerDied","Data":"82b26ab9f9f702629d2a4d98ff4f407ab6fa7b25c433e7fea0c4f8d3edb03d22"} Dec 03 16:06:23 crc kubenswrapper[4998]: I1203 16:06:23.917610 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-49jvv" event={"ID":"69fee984-b61f-4a09-b330-c4aee6a2e41b","Type":"ContainerDied","Data":"f094409fd5dc554e89d534706e20c2ca9eab0e24329d6c6a7b3eb31f074c5ead"} Dec 03 16:06:23 crc kubenswrapper[4998]: I1203 16:06:23.917630 4998 scope.go:117] "RemoveContainer" containerID="82b26ab9f9f702629d2a4d98ff4f407ab6fa7b25c433e7fea0c4f8d3edb03d22" Dec 03 16:06:23 crc kubenswrapper[4998]: I1203 16:06:23.932595 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-49jvv"] Dec 03 16:06:23 crc kubenswrapper[4998]: I1203 16:06:23.935997 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-49jvv"] Dec 03 16:06:23 crc kubenswrapper[4998]: I1203 16:06:23.939041 4998 scope.go:117] "RemoveContainer" containerID="14b61b88c5a5cfed8f79f27cf66ce7484e02b65e19279a0a305e28ca2fd90c3b" Dec 03 16:06:23 crc kubenswrapper[4998]: I1203 16:06:23.952987 4998 scope.go:117] "RemoveContainer" containerID="98b086d9e385eab955edf149d812462ba6899cd9d24b7b00eb4900e435000798" Dec 03 16:06:23 crc kubenswrapper[4998]: I1203 16:06:23.965699 4998 scope.go:117] "RemoveContainer" containerID="82b26ab9f9f702629d2a4d98ff4f407ab6fa7b25c433e7fea0c4f8d3edb03d22" Dec 03 16:06:23 crc kubenswrapper[4998]: E1203 16:06:23.966192 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82b26ab9f9f702629d2a4d98ff4f407ab6fa7b25c433e7fea0c4f8d3edb03d22\": container with ID starting with 82b26ab9f9f702629d2a4d98ff4f407ab6fa7b25c433e7fea0c4f8d3edb03d22 not found: ID does not exist" containerID="82b26ab9f9f702629d2a4d98ff4f407ab6fa7b25c433e7fea0c4f8d3edb03d22" Dec 03 16:06:23 crc kubenswrapper[4998]: I1203 16:06:23.966232 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82b26ab9f9f702629d2a4d98ff4f407ab6fa7b25c433e7fea0c4f8d3edb03d22"} err="failed to get container status \"82b26ab9f9f702629d2a4d98ff4f407ab6fa7b25c433e7fea0c4f8d3edb03d22\": rpc error: code = NotFound desc = could not find container \"82b26ab9f9f702629d2a4d98ff4f407ab6fa7b25c433e7fea0c4f8d3edb03d22\": container with ID starting with 82b26ab9f9f702629d2a4d98ff4f407ab6fa7b25c433e7fea0c4f8d3edb03d22 not found: ID does not exist" Dec 03 16:06:23 crc kubenswrapper[4998]: I1203 16:06:23.966281 4998 scope.go:117] "RemoveContainer" containerID="14b61b88c5a5cfed8f79f27cf66ce7484e02b65e19279a0a305e28ca2fd90c3b" Dec 03 16:06:23 crc kubenswrapper[4998]: E1203 16:06:23.966595 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14b61b88c5a5cfed8f79f27cf66ce7484e02b65e19279a0a305e28ca2fd90c3b\": container with ID starting with 14b61b88c5a5cfed8f79f27cf66ce7484e02b65e19279a0a305e28ca2fd90c3b not found: ID does not exist" containerID="14b61b88c5a5cfed8f79f27cf66ce7484e02b65e19279a0a305e28ca2fd90c3b" Dec 03 16:06:23 crc kubenswrapper[4998]: I1203 16:06:23.966626 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14b61b88c5a5cfed8f79f27cf66ce7484e02b65e19279a0a305e28ca2fd90c3b"} err="failed to get container status \"14b61b88c5a5cfed8f79f27cf66ce7484e02b65e19279a0a305e28ca2fd90c3b\": rpc error: code = NotFound desc = could not find container \"14b61b88c5a5cfed8f79f27cf66ce7484e02b65e19279a0a305e28ca2fd90c3b\": container with ID starting with 14b61b88c5a5cfed8f79f27cf66ce7484e02b65e19279a0a305e28ca2fd90c3b not found: ID does not exist" Dec 03 16:06:23 crc kubenswrapper[4998]: I1203 16:06:23.966652 4998 scope.go:117] "RemoveContainer" containerID="98b086d9e385eab955edf149d812462ba6899cd9d24b7b00eb4900e435000798" Dec 03 16:06:23 crc kubenswrapper[4998]: E1203 16:06:23.967065 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98b086d9e385eab955edf149d812462ba6899cd9d24b7b00eb4900e435000798\": container with ID starting with 98b086d9e385eab955edf149d812462ba6899cd9d24b7b00eb4900e435000798 not found: ID does not exist" containerID="98b086d9e385eab955edf149d812462ba6899cd9d24b7b00eb4900e435000798" Dec 03 16:06:23 crc kubenswrapper[4998]: I1203 16:06:23.967090 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98b086d9e385eab955edf149d812462ba6899cd9d24b7b00eb4900e435000798"} err="failed to get container status \"98b086d9e385eab955edf149d812462ba6899cd9d24b7b00eb4900e435000798\": rpc error: code = NotFound desc = could not find container \"98b086d9e385eab955edf149d812462ba6899cd9d24b7b00eb4900e435000798\": container with ID starting with 98b086d9e385eab955edf149d812462ba6899cd9d24b7b00eb4900e435000798 not found: ID does not exist" Dec 03 16:06:25 crc kubenswrapper[4998]: I1203 16:06:25.685889 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69fee984-b61f-4a09-b330-c4aee6a2e41b" path="/var/lib/kubelet/pods/69fee984-b61f-4a09-b330-c4aee6a2e41b/volumes" Dec 03 16:06:27 crc kubenswrapper[4998]: I1203 16:06:27.111285 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:06:27 crc kubenswrapper[4998]: I1203 16:06:27.112053 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:06:27 crc kubenswrapper[4998]: I1203 16:06:27.345390 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-84rhg" Dec 03 16:06:27 crc kubenswrapper[4998]: I1203 16:06:27.345449 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-84rhg" Dec 03 16:06:27 crc kubenswrapper[4998]: I1203 16:06:27.757875 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-czz6d" Dec 03 16:06:27 crc kubenswrapper[4998]: I1203 16:06:27.758247 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-czz6d" Dec 03 16:06:27 crc kubenswrapper[4998]: I1203 16:06:27.797403 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-czz6d" Dec 03 16:06:27 crc kubenswrapper[4998]: I1203 16:06:27.940386 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5v5j8" Dec 03 16:06:27 crc kubenswrapper[4998]: I1203 16:06:27.940455 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5v5j8" Dec 03 16:06:27 crc kubenswrapper[4998]: I1203 16:06:27.984492 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5v5j8" Dec 03 16:06:28 crc kubenswrapper[4998]: I1203 16:06:28.415612 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-84rhg" podUID="8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea" containerName="registry-server" probeResult="failure" output=< Dec 03 16:06:28 crc kubenswrapper[4998]: timeout: failed to connect service ":50051" within 1s Dec 03 16:06:28 crc kubenswrapper[4998]: > Dec 03 16:06:28 crc kubenswrapper[4998]: I1203 16:06:28.577910 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/community-operators-rrdtb" podUID="5ef027be-08d9-4731-bd35-34f7bb2fe43a" containerName="registry-server" probeResult="failure" output=< Dec 03 16:06:28 crc kubenswrapper[4998]: timeout: failed to connect service ":50051" within 1s Dec 03 16:06:28 crc kubenswrapper[4998]: > Dec 03 16:06:28 crc kubenswrapper[4998]: I1203 16:06:28.582942 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/community-operators-rrdtb" podUID="5ef027be-08d9-4731-bd35-34f7bb2fe43a" containerName="registry-server" probeResult="failure" output=< Dec 03 16:06:28 crc kubenswrapper[4998]: timeout: failed to connect service ":50051" within 1s Dec 03 16:06:28 crc kubenswrapper[4998]: > Dec 03 16:06:28 crc kubenswrapper[4998]: I1203 16:06:28.751365 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-czz6d" Dec 03 16:06:28 crc kubenswrapper[4998]: I1203 16:06:28.984355 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5v5j8" Dec 03 16:06:29 crc kubenswrapper[4998]: I1203 16:06:29.621460 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5v5j8"] Dec 03 16:06:29 crc kubenswrapper[4998]: I1203 16:06:29.763125 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dsh8q" Dec 03 16:06:29 crc kubenswrapper[4998]: I1203 16:06:29.763448 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dsh8q" Dec 03 16:06:29 crc kubenswrapper[4998]: I1203 16:06:29.816673 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dsh8q" Dec 03 16:06:30 crc kubenswrapper[4998]: I1203 16:06:30.007482 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dsh8q" Dec 03 16:06:30 crc kubenswrapper[4998]: I1203 16:06:30.138059 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c7jnc" Dec 03 16:06:30 crc kubenswrapper[4998]: I1203 16:06:30.138337 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c7jnc" Dec 03 16:06:30 crc kubenswrapper[4998]: I1203 16:06:30.172921 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c7jnc" Dec 03 16:06:30 crc kubenswrapper[4998]: I1203 16:06:30.554510 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6rx7t" Dec 03 16:06:30 crc kubenswrapper[4998]: I1203 16:06:30.554589 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6rx7t" Dec 03 16:06:30 crc kubenswrapper[4998]: I1203 16:06:30.613463 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6rx7t" Dec 03 16:06:30 crc kubenswrapper[4998]: I1203 16:06:30.958536 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5v5j8" podUID="97fb2ff7-4894-4858-99aa-1a54a3119d33" containerName="registry-server" containerID="cri-o://570e5c73768e5818016141650b8627643e932b16aa0ef1e5bad262256a9f109d" gracePeriod=2 Dec 03 16:06:31 crc kubenswrapper[4998]: I1203 16:06:31.003432 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6rx7t" Dec 03 16:06:31 crc kubenswrapper[4998]: I1203 16:06:31.015829 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c7jnc" Dec 03 16:06:31 crc kubenswrapper[4998]: I1203 16:06:31.019085 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-czz6d"] Dec 03 16:06:31 crc kubenswrapper[4998]: I1203 16:06:31.019345 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-czz6d" podUID="50083a46-5fd2-4241-8a20-67ae6cd5f6b6" containerName="registry-server" containerID="cri-o://1d31a6843d11f0c77a603504c4bfc4429fda271c8c4fbc39a9e25a04a012d09c" gracePeriod=2 Dec 03 16:06:33 crc kubenswrapper[4998]: I1203 16:06:33.422446 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c7jnc"] Dec 03 16:06:33 crc kubenswrapper[4998]: I1203 16:06:33.978089 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c7jnc" podUID="d66602dc-ef95-4eec-989f-8e42dc4bcd02" containerName="registry-server" containerID="cri-o://a9a5ccb5aff999ee23d08c44e58d8f36ce41d467a1ba4d5163add64d310f1c95" gracePeriod=2 Dec 03 16:06:37 crc kubenswrapper[4998]: I1203 16:06:37.409637 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-84rhg" Dec 03 16:06:37 crc kubenswrapper[4998]: I1203 16:06:37.480741 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-84rhg" Dec 03 16:06:37 crc kubenswrapper[4998]: E1203 16:06:37.758739 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1d31a6843d11f0c77a603504c4bfc4429fda271c8c4fbc39a9e25a04a012d09c is running failed: container process not found" containerID="1d31a6843d11f0c77a603504c4bfc4429fda271c8c4fbc39a9e25a04a012d09c" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 16:06:37 crc kubenswrapper[4998]: E1203 16:06:37.759414 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1d31a6843d11f0c77a603504c4bfc4429fda271c8c4fbc39a9e25a04a012d09c is running failed: container process not found" containerID="1d31a6843d11f0c77a603504c4bfc4429fda271c8c4fbc39a9e25a04a012d09c" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 16:06:37 crc kubenswrapper[4998]: E1203 16:06:37.760207 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1d31a6843d11f0c77a603504c4bfc4429fda271c8c4fbc39a9e25a04a012d09c is running failed: container process not found" containerID="1d31a6843d11f0c77a603504c4bfc4429fda271c8c4fbc39a9e25a04a012d09c" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 16:06:37 crc kubenswrapper[4998]: E1203 16:06:37.760291 4998 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1d31a6843d11f0c77a603504c4bfc4429fda271c8c4fbc39a9e25a04a012d09c is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-czz6d" podUID="50083a46-5fd2-4241-8a20-67ae6cd5f6b6" containerName="registry-server" Dec 03 16:06:37 crc kubenswrapper[4998]: E1203 16:06:37.941511 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 570e5c73768e5818016141650b8627643e932b16aa0ef1e5bad262256a9f109d is running failed: container process not found" containerID="570e5c73768e5818016141650b8627643e932b16aa0ef1e5bad262256a9f109d" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 16:06:37 crc kubenswrapper[4998]: E1203 16:06:37.942309 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 570e5c73768e5818016141650b8627643e932b16aa0ef1e5bad262256a9f109d is running failed: container process not found" containerID="570e5c73768e5818016141650b8627643e932b16aa0ef1e5bad262256a9f109d" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 16:06:37 crc kubenswrapper[4998]: E1203 16:06:37.943068 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 570e5c73768e5818016141650b8627643e932b16aa0ef1e5bad262256a9f109d is running failed: container process not found" containerID="570e5c73768e5818016141650b8627643e932b16aa0ef1e5bad262256a9f109d" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 16:06:37 crc kubenswrapper[4998]: E1203 16:06:37.943142 4998 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 570e5c73768e5818016141650b8627643e932b16aa0ef1e5bad262256a9f109d is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-5v5j8" podUID="97fb2ff7-4894-4858-99aa-1a54a3119d33" containerName="registry-server" Dec 03 16:06:38 crc kubenswrapper[4998]: I1203 16:06:38.702724 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" podUID="56ad3380-4861-49d1-8758-7b1e27f74560" containerName="oauth-openshift" containerID="cri-o://4cfc40eef4ee0dce61eb0b953949c62810ea4c11b954526d5e8450ec31a260fa" gracePeriod=15 Dec 03 16:06:39 crc kubenswrapper[4998]: I1203 16:06:39.694233 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5v5j8_97fb2ff7-4894-4858-99aa-1a54a3119d33/registry-server/0.log" Dec 03 16:06:39 crc kubenswrapper[4998]: I1203 16:06:39.695346 4998 generic.go:334] "Generic (PLEG): container finished" podID="97fb2ff7-4894-4858-99aa-1a54a3119d33" containerID="570e5c73768e5818016141650b8627643e932b16aa0ef1e5bad262256a9f109d" exitCode=137 Dec 03 16:06:39 crc kubenswrapper[4998]: I1203 16:06:39.695420 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5v5j8" event={"ID":"97fb2ff7-4894-4858-99aa-1a54a3119d33","Type":"ContainerDied","Data":"570e5c73768e5818016141650b8627643e932b16aa0ef1e5bad262256a9f109d"} Dec 03 16:06:39 crc kubenswrapper[4998]: I1203 16:06:39.697300 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-czz6d_50083a46-5fd2-4241-8a20-67ae6cd5f6b6/registry-server/0.log" Dec 03 16:06:39 crc kubenswrapper[4998]: I1203 16:06:39.697981 4998 generic.go:334] "Generic (PLEG): container finished" podID="50083a46-5fd2-4241-8a20-67ae6cd5f6b6" containerID="1d31a6843d11f0c77a603504c4bfc4429fda271c8c4fbc39a9e25a04a012d09c" exitCode=137 Dec 03 16:06:39 crc kubenswrapper[4998]: I1203 16:06:39.698017 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-czz6d" event={"ID":"50083a46-5fd2-4241-8a20-67ae6cd5f6b6","Type":"ContainerDied","Data":"1d31a6843d11f0c77a603504c4bfc4429fda271c8c4fbc39a9e25a04a012d09c"} Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.025965 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-czz6d_50083a46-5fd2-4241-8a20-67ae6cd5f6b6/registry-server/0.log" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.026876 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-czz6d" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.117888 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrvbv\" (UniqueName: \"kubernetes.io/projected/50083a46-5fd2-4241-8a20-67ae6cd5f6b6-kube-api-access-vrvbv\") pod \"50083a46-5fd2-4241-8a20-67ae6cd5f6b6\" (UID: \"50083a46-5fd2-4241-8a20-67ae6cd5f6b6\") " Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.118062 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50083a46-5fd2-4241-8a20-67ae6cd5f6b6-catalog-content\") pod \"50083a46-5fd2-4241-8a20-67ae6cd5f6b6\" (UID: \"50083a46-5fd2-4241-8a20-67ae6cd5f6b6\") " Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.118157 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50083a46-5fd2-4241-8a20-67ae6cd5f6b6-utilities\") pod \"50083a46-5fd2-4241-8a20-67ae6cd5f6b6\" (UID: \"50083a46-5fd2-4241-8a20-67ae6cd5f6b6\") " Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.119393 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50083a46-5fd2-4241-8a20-67ae6cd5f6b6-utilities" (OuterVolumeSpecName: "utilities") pod "50083a46-5fd2-4241-8a20-67ae6cd5f6b6" (UID: "50083a46-5fd2-4241-8a20-67ae6cd5f6b6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.129133 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50083a46-5fd2-4241-8a20-67ae6cd5f6b6-kube-api-access-vrvbv" (OuterVolumeSpecName: "kube-api-access-vrvbv") pod "50083a46-5fd2-4241-8a20-67ae6cd5f6b6" (UID: "50083a46-5fd2-4241-8a20-67ae6cd5f6b6"). InnerVolumeSpecName "kube-api-access-vrvbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:06:40 crc kubenswrapper[4998]: E1203 16:06:40.138435 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a9a5ccb5aff999ee23d08c44e58d8f36ce41d467a1ba4d5163add64d310f1c95 is running failed: container process not found" containerID="a9a5ccb5aff999ee23d08c44e58d8f36ce41d467a1ba4d5163add64d310f1c95" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 16:06:40 crc kubenswrapper[4998]: E1203 16:06:40.138807 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a9a5ccb5aff999ee23d08c44e58d8f36ce41d467a1ba4d5163add64d310f1c95 is running failed: container process not found" containerID="a9a5ccb5aff999ee23d08c44e58d8f36ce41d467a1ba4d5163add64d310f1c95" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 16:06:40 crc kubenswrapper[4998]: E1203 16:06:40.139127 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a9a5ccb5aff999ee23d08c44e58d8f36ce41d467a1ba4d5163add64d310f1c95 is running failed: container process not found" containerID="a9a5ccb5aff999ee23d08c44e58d8f36ce41d467a1ba4d5163add64d310f1c95" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 16:06:40 crc kubenswrapper[4998]: E1203 16:06:40.139175 4998 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a9a5ccb5aff999ee23d08c44e58d8f36ce41d467a1ba4d5163add64d310f1c95 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-c7jnc" podUID="d66602dc-ef95-4eec-989f-8e42dc4bcd02" containerName="registry-server" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.152724 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c7jnc" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.188826 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50083a46-5fd2-4241-8a20-67ae6cd5f6b6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "50083a46-5fd2-4241-8a20-67ae6cd5f6b6" (UID: "50083a46-5fd2-4241-8a20-67ae6cd5f6b6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.198529 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5v5j8_97fb2ff7-4894-4858-99aa-1a54a3119d33/registry-server/0.log" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.199697 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5v5j8" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.219342 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwgsn\" (UniqueName: \"kubernetes.io/projected/d66602dc-ef95-4eec-989f-8e42dc4bcd02-kube-api-access-kwgsn\") pod \"d66602dc-ef95-4eec-989f-8e42dc4bcd02\" (UID: \"d66602dc-ef95-4eec-989f-8e42dc4bcd02\") " Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.219383 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d66602dc-ef95-4eec-989f-8e42dc4bcd02-utilities\") pod \"d66602dc-ef95-4eec-989f-8e42dc4bcd02\" (UID: \"d66602dc-ef95-4eec-989f-8e42dc4bcd02\") " Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.219420 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d66602dc-ef95-4eec-989f-8e42dc4bcd02-catalog-content\") pod \"d66602dc-ef95-4eec-989f-8e42dc4bcd02\" (UID: \"d66602dc-ef95-4eec-989f-8e42dc4bcd02\") " Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.219819 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50083a46-5fd2-4241-8a20-67ae6cd5f6b6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.219836 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50083a46-5fd2-4241-8a20-67ae6cd5f6b6-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.219848 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrvbv\" (UniqueName: \"kubernetes.io/projected/50083a46-5fd2-4241-8a20-67ae6cd5f6b6-kube-api-access-vrvbv\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.225007 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d66602dc-ef95-4eec-989f-8e42dc4bcd02-kube-api-access-kwgsn" (OuterVolumeSpecName: "kube-api-access-kwgsn") pod "d66602dc-ef95-4eec-989f-8e42dc4bcd02" (UID: "d66602dc-ef95-4eec-989f-8e42dc4bcd02"). InnerVolumeSpecName "kube-api-access-kwgsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.226911 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d66602dc-ef95-4eec-989f-8e42dc4bcd02-utilities" (OuterVolumeSpecName: "utilities") pod "d66602dc-ef95-4eec-989f-8e42dc4bcd02" (UID: "d66602dc-ef95-4eec-989f-8e42dc4bcd02"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.228237 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.238409 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d66602dc-ef95-4eec-989f-8e42dc4bcd02-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d66602dc-ef95-4eec-989f-8e42dc4bcd02" (UID: "d66602dc-ef95-4eec-989f-8e42dc4bcd02"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.321411 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97fb2ff7-4894-4858-99aa-1a54a3119d33-catalog-content\") pod \"97fb2ff7-4894-4858-99aa-1a54a3119d33\" (UID: \"97fb2ff7-4894-4858-99aa-1a54a3119d33\") " Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.321487 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-cliconfig\") pod \"56ad3380-4861-49d1-8758-7b1e27f74560\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.321524 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-serving-cert\") pod \"56ad3380-4861-49d1-8758-7b1e27f74560\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.321618 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-user-template-provider-selection\") pod \"56ad3380-4861-49d1-8758-7b1e27f74560\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.321654 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2r7z\" (UniqueName: \"kubernetes.io/projected/56ad3380-4861-49d1-8758-7b1e27f74560-kube-api-access-n2r7z\") pod \"56ad3380-4861-49d1-8758-7b1e27f74560\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.321690 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-session\") pod \"56ad3380-4861-49d1-8758-7b1e27f74560\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.321730 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-user-idp-0-file-data\") pod \"56ad3380-4861-49d1-8758-7b1e27f74560\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.321789 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-service-ca\") pod \"56ad3380-4861-49d1-8758-7b1e27f74560\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.321821 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97fb2ff7-4894-4858-99aa-1a54a3119d33-utilities\") pod \"97fb2ff7-4894-4858-99aa-1a54a3119d33\" (UID: \"97fb2ff7-4894-4858-99aa-1a54a3119d33\") " Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.321850 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/56ad3380-4861-49d1-8758-7b1e27f74560-audit-dir\") pod \"56ad3380-4861-49d1-8758-7b1e27f74560\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.321884 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-trusted-ca-bundle\") pod \"56ad3380-4861-49d1-8758-7b1e27f74560\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.321935 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/56ad3380-4861-49d1-8758-7b1e27f74560-audit-policies\") pod \"56ad3380-4861-49d1-8758-7b1e27f74560\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.321975 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-router-certs\") pod \"56ad3380-4861-49d1-8758-7b1e27f74560\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.322020 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-user-template-error\") pod \"56ad3380-4861-49d1-8758-7b1e27f74560\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.322059 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-ocp-branding-template\") pod \"56ad3380-4861-49d1-8758-7b1e27f74560\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.322092 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-user-template-login\") pod \"56ad3380-4861-49d1-8758-7b1e27f74560\" (UID: \"56ad3380-4861-49d1-8758-7b1e27f74560\") " Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.322139 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcgtc\" (UniqueName: \"kubernetes.io/projected/97fb2ff7-4894-4858-99aa-1a54a3119d33-kube-api-access-hcgtc\") pod \"97fb2ff7-4894-4858-99aa-1a54a3119d33\" (UID: \"97fb2ff7-4894-4858-99aa-1a54a3119d33\") " Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.322295 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "56ad3380-4861-49d1-8758-7b1e27f74560" (UID: "56ad3380-4861-49d1-8758-7b1e27f74560"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.322381 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/56ad3380-4861-49d1-8758-7b1e27f74560-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "56ad3380-4861-49d1-8758-7b1e27f74560" (UID: "56ad3380-4861-49d1-8758-7b1e27f74560"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.322724 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwgsn\" (UniqueName: \"kubernetes.io/projected/d66602dc-ef95-4eec-989f-8e42dc4bcd02-kube-api-access-kwgsn\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.322772 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d66602dc-ef95-4eec-989f-8e42dc4bcd02-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.322821 4998 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/56ad3380-4861-49d1-8758-7b1e27f74560-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.322842 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d66602dc-ef95-4eec-989f-8e42dc4bcd02-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.322859 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "56ad3380-4861-49d1-8758-7b1e27f74560" (UID: "56ad3380-4861-49d1-8758-7b1e27f74560"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.322859 4998 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.323188 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97fb2ff7-4894-4858-99aa-1a54a3119d33-utilities" (OuterVolumeSpecName: "utilities") pod "97fb2ff7-4894-4858-99aa-1a54a3119d33" (UID: "97fb2ff7-4894-4858-99aa-1a54a3119d33"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.324002 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56ad3380-4861-49d1-8758-7b1e27f74560-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "56ad3380-4861-49d1-8758-7b1e27f74560" (UID: "56ad3380-4861-49d1-8758-7b1e27f74560"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.324556 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "56ad3380-4861-49d1-8758-7b1e27f74560" (UID: "56ad3380-4861-49d1-8758-7b1e27f74560"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.325052 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "56ad3380-4861-49d1-8758-7b1e27f74560" (UID: "56ad3380-4861-49d1-8758-7b1e27f74560"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.325334 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "56ad3380-4861-49d1-8758-7b1e27f74560" (UID: "56ad3380-4861-49d1-8758-7b1e27f74560"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.326117 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "56ad3380-4861-49d1-8758-7b1e27f74560" (UID: "56ad3380-4861-49d1-8758-7b1e27f74560"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.326880 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "56ad3380-4861-49d1-8758-7b1e27f74560" (UID: "56ad3380-4861-49d1-8758-7b1e27f74560"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.326896 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97fb2ff7-4894-4858-99aa-1a54a3119d33-kube-api-access-hcgtc" (OuterVolumeSpecName: "kube-api-access-hcgtc") pod "97fb2ff7-4894-4858-99aa-1a54a3119d33" (UID: "97fb2ff7-4894-4858-99aa-1a54a3119d33"). InnerVolumeSpecName "kube-api-access-hcgtc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.327174 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "56ad3380-4861-49d1-8758-7b1e27f74560" (UID: "56ad3380-4861-49d1-8758-7b1e27f74560"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.327394 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "56ad3380-4861-49d1-8758-7b1e27f74560" (UID: "56ad3380-4861-49d1-8758-7b1e27f74560"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.327476 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "56ad3380-4861-49d1-8758-7b1e27f74560" (UID: "56ad3380-4861-49d1-8758-7b1e27f74560"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.328375 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56ad3380-4861-49d1-8758-7b1e27f74560-kube-api-access-n2r7z" (OuterVolumeSpecName: "kube-api-access-n2r7z") pod "56ad3380-4861-49d1-8758-7b1e27f74560" (UID: "56ad3380-4861-49d1-8758-7b1e27f74560"). InnerVolumeSpecName "kube-api-access-n2r7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.328915 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "56ad3380-4861-49d1-8758-7b1e27f74560" (UID: "56ad3380-4861-49d1-8758-7b1e27f74560"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.388417 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97fb2ff7-4894-4858-99aa-1a54a3119d33-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "97fb2ff7-4894-4858-99aa-1a54a3119d33" (UID: "97fb2ff7-4894-4858-99aa-1a54a3119d33"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.425193 4998 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.425249 4998 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.425312 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2r7z\" (UniqueName: \"kubernetes.io/projected/56ad3380-4861-49d1-8758-7b1e27f74560-kube-api-access-n2r7z\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.425331 4998 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.425388 4998 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.425408 4998 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.425427 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97fb2ff7-4894-4858-99aa-1a54a3119d33-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.425482 4998 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.425503 4998 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/56ad3380-4861-49d1-8758-7b1e27f74560-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.425521 4998 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.425541 4998 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.425563 4998 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.425584 4998 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/56ad3380-4861-49d1-8758-7b1e27f74560-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.425607 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcgtc\" (UniqueName: \"kubernetes.io/projected/97fb2ff7-4894-4858-99aa-1a54a3119d33-kube-api-access-hcgtc\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.425624 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97fb2ff7-4894-4858-99aa-1a54a3119d33-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.706356 4998 generic.go:334] "Generic (PLEG): container finished" podID="56ad3380-4861-49d1-8758-7b1e27f74560" containerID="4cfc40eef4ee0dce61eb0b953949c62810ea4c11b954526d5e8450ec31a260fa" exitCode=0 Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.706417 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.706442 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" event={"ID":"56ad3380-4861-49d1-8758-7b1e27f74560","Type":"ContainerDied","Data":"4cfc40eef4ee0dce61eb0b953949c62810ea4c11b954526d5e8450ec31a260fa"} Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.706493 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-b5hpf" event={"ID":"56ad3380-4861-49d1-8758-7b1e27f74560","Type":"ContainerDied","Data":"f6b1e607b8be12fd2922ddb6c59e52604deb2fdad5903c35d87e7586f4f82867"} Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.706624 4998 scope.go:117] "RemoveContainer" containerID="4cfc40eef4ee0dce61eb0b953949c62810ea4c11b954526d5e8450ec31a260fa" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.710665 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-czz6d_50083a46-5fd2-4241-8a20-67ae6cd5f6b6/registry-server/0.log" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.712952 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-czz6d" event={"ID":"50083a46-5fd2-4241-8a20-67ae6cd5f6b6","Type":"ContainerDied","Data":"6fe6b091ba814eb11f5421dca011d5d93fa0bfeba4c41534c59e756245cb8913"} Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.713094 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-czz6d" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.718356 4998 generic.go:334] "Generic (PLEG): container finished" podID="d66602dc-ef95-4eec-989f-8e42dc4bcd02" containerID="a9a5ccb5aff999ee23d08c44e58d8f36ce41d467a1ba4d5163add64d310f1c95" exitCode=0 Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.718436 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7jnc" event={"ID":"d66602dc-ef95-4eec-989f-8e42dc4bcd02","Type":"ContainerDied","Data":"a9a5ccb5aff999ee23d08c44e58d8f36ce41d467a1ba4d5163add64d310f1c95"} Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.718469 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c7jnc" event={"ID":"d66602dc-ef95-4eec-989f-8e42dc4bcd02","Type":"ContainerDied","Data":"d6c4307c2f25ad7381a0a4105491285868587d480028b7d4d668ee9526a3fc6b"} Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.718555 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c7jnc" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.734633 4998 scope.go:117] "RemoveContainer" containerID="4cfc40eef4ee0dce61eb0b953949c62810ea4c11b954526d5e8450ec31a260fa" Dec 03 16:06:40 crc kubenswrapper[4998]: E1203 16:06:40.737321 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cfc40eef4ee0dce61eb0b953949c62810ea4c11b954526d5e8450ec31a260fa\": container with ID starting with 4cfc40eef4ee0dce61eb0b953949c62810ea4c11b954526d5e8450ec31a260fa not found: ID does not exist" containerID="4cfc40eef4ee0dce61eb0b953949c62810ea4c11b954526d5e8450ec31a260fa" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.737376 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cfc40eef4ee0dce61eb0b953949c62810ea4c11b954526d5e8450ec31a260fa"} err="failed to get container status \"4cfc40eef4ee0dce61eb0b953949c62810ea4c11b954526d5e8450ec31a260fa\": rpc error: code = NotFound desc = could not find container \"4cfc40eef4ee0dce61eb0b953949c62810ea4c11b954526d5e8450ec31a260fa\": container with ID starting with 4cfc40eef4ee0dce61eb0b953949c62810ea4c11b954526d5e8450ec31a260fa not found: ID does not exist" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.737411 4998 scope.go:117] "RemoveContainer" containerID="1d31a6843d11f0c77a603504c4bfc4429fda271c8c4fbc39a9e25a04a012d09c" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.737771 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5v5j8_97fb2ff7-4894-4858-99aa-1a54a3119d33/registry-server/0.log" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.744049 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5v5j8" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.744111 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5v5j8" event={"ID":"97fb2ff7-4894-4858-99aa-1a54a3119d33","Type":"ContainerDied","Data":"12cde948b68a3f0837ca6d4e5518e0920bf4e842cb029d1cf8c9e3459c022eb1"} Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.768127 4998 scope.go:117] "RemoveContainer" containerID="d96a0239f749b70416e1f84fb2d6064e6d9c4413642bb00009e4984e08d80da5" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.772815 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-czz6d"] Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.778178 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-czz6d"] Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.783231 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-b5hpf"] Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.786991 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-b5hpf"] Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.798211 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c7jnc"] Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.803800 4998 scope.go:117] "RemoveContainer" containerID="d29809dcd5982fb5233b79073def5e2ea6290e274c54d747abec4bf91e31ed27" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.806071 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c7jnc"] Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.816332 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5v5j8"] Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.821889 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5v5j8"] Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.832024 4998 scope.go:117] "RemoveContainer" containerID="a9a5ccb5aff999ee23d08c44e58d8f36ce41d467a1ba4d5163add64d310f1c95" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.847560 4998 scope.go:117] "RemoveContainer" containerID="7e34c2ae31325ba2b7df3c95a9afd1a7e806e78ad412fa849dd99df19c6155b5" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.864661 4998 scope.go:117] "RemoveContainer" containerID="745579b66f1f14386edc835f36b411aad0b996a24fd1999a2116d23fa8a71646" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.878358 4998 scope.go:117] "RemoveContainer" containerID="a9a5ccb5aff999ee23d08c44e58d8f36ce41d467a1ba4d5163add64d310f1c95" Dec 03 16:06:40 crc kubenswrapper[4998]: E1203 16:06:40.879011 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9a5ccb5aff999ee23d08c44e58d8f36ce41d467a1ba4d5163add64d310f1c95\": container with ID starting with a9a5ccb5aff999ee23d08c44e58d8f36ce41d467a1ba4d5163add64d310f1c95 not found: ID does not exist" containerID="a9a5ccb5aff999ee23d08c44e58d8f36ce41d467a1ba4d5163add64d310f1c95" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.879048 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9a5ccb5aff999ee23d08c44e58d8f36ce41d467a1ba4d5163add64d310f1c95"} err="failed to get container status \"a9a5ccb5aff999ee23d08c44e58d8f36ce41d467a1ba4d5163add64d310f1c95\": rpc error: code = NotFound desc = could not find container \"a9a5ccb5aff999ee23d08c44e58d8f36ce41d467a1ba4d5163add64d310f1c95\": container with ID starting with a9a5ccb5aff999ee23d08c44e58d8f36ce41d467a1ba4d5163add64d310f1c95 not found: ID does not exist" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.879079 4998 scope.go:117] "RemoveContainer" containerID="7e34c2ae31325ba2b7df3c95a9afd1a7e806e78ad412fa849dd99df19c6155b5" Dec 03 16:06:40 crc kubenswrapper[4998]: E1203 16:06:40.879624 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e34c2ae31325ba2b7df3c95a9afd1a7e806e78ad412fa849dd99df19c6155b5\": container with ID starting with 7e34c2ae31325ba2b7df3c95a9afd1a7e806e78ad412fa849dd99df19c6155b5 not found: ID does not exist" containerID="7e34c2ae31325ba2b7df3c95a9afd1a7e806e78ad412fa849dd99df19c6155b5" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.879656 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e34c2ae31325ba2b7df3c95a9afd1a7e806e78ad412fa849dd99df19c6155b5"} err="failed to get container status \"7e34c2ae31325ba2b7df3c95a9afd1a7e806e78ad412fa849dd99df19c6155b5\": rpc error: code = NotFound desc = could not find container \"7e34c2ae31325ba2b7df3c95a9afd1a7e806e78ad412fa849dd99df19c6155b5\": container with ID starting with 7e34c2ae31325ba2b7df3c95a9afd1a7e806e78ad412fa849dd99df19c6155b5 not found: ID does not exist" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.879676 4998 scope.go:117] "RemoveContainer" containerID="745579b66f1f14386edc835f36b411aad0b996a24fd1999a2116d23fa8a71646" Dec 03 16:06:40 crc kubenswrapper[4998]: E1203 16:06:40.880189 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"745579b66f1f14386edc835f36b411aad0b996a24fd1999a2116d23fa8a71646\": container with ID starting with 745579b66f1f14386edc835f36b411aad0b996a24fd1999a2116d23fa8a71646 not found: ID does not exist" containerID="745579b66f1f14386edc835f36b411aad0b996a24fd1999a2116d23fa8a71646" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.880248 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"745579b66f1f14386edc835f36b411aad0b996a24fd1999a2116d23fa8a71646"} err="failed to get container status \"745579b66f1f14386edc835f36b411aad0b996a24fd1999a2116d23fa8a71646\": rpc error: code = NotFound desc = could not find container \"745579b66f1f14386edc835f36b411aad0b996a24fd1999a2116d23fa8a71646\": container with ID starting with 745579b66f1f14386edc835f36b411aad0b996a24fd1999a2116d23fa8a71646 not found: ID does not exist" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.880295 4998 scope.go:117] "RemoveContainer" containerID="570e5c73768e5818016141650b8627643e932b16aa0ef1e5bad262256a9f109d" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.894955 4998 scope.go:117] "RemoveContainer" containerID="5a2df0bd563fc1e5507c165e109d53270e713b31b63e9f1388677886d0e20d86" Dec 03 16:06:40 crc kubenswrapper[4998]: I1203 16:06:40.921128 4998 scope.go:117] "RemoveContainer" containerID="2d44e443563c313d23cad9b53b4cad1fbdec714b71d6264fcae19bc4db5281b0" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.418018 4998 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 16:06:41 crc kubenswrapper[4998]: E1203 16:06:41.419381 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97fb2ff7-4894-4858-99aa-1a54a3119d33" containerName="extract-utilities" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.419489 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="97fb2ff7-4894-4858-99aa-1a54a3119d33" containerName="extract-utilities" Dec 03 16:06:41 crc kubenswrapper[4998]: E1203 16:06:41.419578 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69fee984-b61f-4a09-b330-c4aee6a2e41b" containerName="extract-utilities" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.419661 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="69fee984-b61f-4a09-b330-c4aee6a2e41b" containerName="extract-utilities" Dec 03 16:06:41 crc kubenswrapper[4998]: E1203 16:06:41.419771 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97fb2ff7-4894-4858-99aa-1a54a3119d33" containerName="extract-content" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.419860 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="97fb2ff7-4894-4858-99aa-1a54a3119d33" containerName="extract-content" Dec 03 16:06:41 crc kubenswrapper[4998]: E1203 16:06:41.419948 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69fee984-b61f-4a09-b330-c4aee6a2e41b" containerName="extract-content" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.420021 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="69fee984-b61f-4a09-b330-c4aee6a2e41b" containerName="extract-content" Dec 03 16:06:41 crc kubenswrapper[4998]: E1203 16:06:41.420108 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50083a46-5fd2-4241-8a20-67ae6cd5f6b6" containerName="extract-utilities" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.420181 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="50083a46-5fd2-4241-8a20-67ae6cd5f6b6" containerName="extract-utilities" Dec 03 16:06:41 crc kubenswrapper[4998]: E1203 16:06:41.420254 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d66602dc-ef95-4eec-989f-8e42dc4bcd02" containerName="extract-utilities" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.420325 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="d66602dc-ef95-4eec-989f-8e42dc4bcd02" containerName="extract-utilities" Dec 03 16:06:41 crc kubenswrapper[4998]: E1203 16:06:41.420406 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50083a46-5fd2-4241-8a20-67ae6cd5f6b6" containerName="extract-content" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.420597 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="50083a46-5fd2-4241-8a20-67ae6cd5f6b6" containerName="extract-content" Dec 03 16:06:41 crc kubenswrapper[4998]: E1203 16:06:41.420682 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56ad3380-4861-49d1-8758-7b1e27f74560" containerName="oauth-openshift" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.420782 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="56ad3380-4861-49d1-8758-7b1e27f74560" containerName="oauth-openshift" Dec 03 16:06:41 crc kubenswrapper[4998]: E1203 16:06:41.420862 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69fee984-b61f-4a09-b330-c4aee6a2e41b" containerName="registry-server" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.420951 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="69fee984-b61f-4a09-b330-c4aee6a2e41b" containerName="registry-server" Dec 03 16:06:41 crc kubenswrapper[4998]: E1203 16:06:41.421029 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97fb2ff7-4894-4858-99aa-1a54a3119d33" containerName="registry-server" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.421108 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="97fb2ff7-4894-4858-99aa-1a54a3119d33" containerName="registry-server" Dec 03 16:06:41 crc kubenswrapper[4998]: E1203 16:06:41.421187 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e4733c6-a2dd-4beb-8240-6e57afff93cb" containerName="pruner" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.421268 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e4733c6-a2dd-4beb-8240-6e57afff93cb" containerName="pruner" Dec 03 16:06:41 crc kubenswrapper[4998]: E1203 16:06:41.421343 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50083a46-5fd2-4241-8a20-67ae6cd5f6b6" containerName="registry-server" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.421414 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="50083a46-5fd2-4241-8a20-67ae6cd5f6b6" containerName="registry-server" Dec 03 16:06:41 crc kubenswrapper[4998]: E1203 16:06:41.421490 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d66602dc-ef95-4eec-989f-8e42dc4bcd02" containerName="registry-server" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.421570 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="d66602dc-ef95-4eec-989f-8e42dc4bcd02" containerName="registry-server" Dec 03 16:06:41 crc kubenswrapper[4998]: E1203 16:06:41.421658 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d66602dc-ef95-4eec-989f-8e42dc4bcd02" containerName="extract-content" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.421735 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="d66602dc-ef95-4eec-989f-8e42dc4bcd02" containerName="extract-content" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.421964 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e4733c6-a2dd-4beb-8240-6e57afff93cb" containerName="pruner" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.422051 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="d66602dc-ef95-4eec-989f-8e42dc4bcd02" containerName="registry-server" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.422130 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="69fee984-b61f-4a09-b330-c4aee6a2e41b" containerName="registry-server" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.422206 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="50083a46-5fd2-4241-8a20-67ae6cd5f6b6" containerName="registry-server" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.422294 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="97fb2ff7-4894-4858-99aa-1a54a3119d33" containerName="registry-server" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.422373 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="56ad3380-4861-49d1-8758-7b1e27f74560" containerName="oauth-openshift" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.422926 4998 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.423116 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.423211 4998 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.423578 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079" gracePeriod=15 Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.423639 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2" gracePeriod=15 Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.423591 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56" gracePeriod=15 Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.423688 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb" gracePeriod=15 Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.423772 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a" gracePeriod=15 Dec 03 16:06:41 crc kubenswrapper[4998]: E1203 16:06:41.424006 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.424045 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 16:06:41 crc kubenswrapper[4998]: E1203 16:06:41.424065 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.424073 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 16:06:41 crc kubenswrapper[4998]: E1203 16:06:41.424092 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.424100 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 16:06:41 crc kubenswrapper[4998]: E1203 16:06:41.424110 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.424118 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 16:06:41 crc kubenswrapper[4998]: E1203 16:06:41.424127 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.424134 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 16:06:41 crc kubenswrapper[4998]: E1203 16:06:41.424149 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.424157 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.424284 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.424300 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.424313 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.424323 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.424331 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.424342 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 16:06:41 crc kubenswrapper[4998]: E1203 16:06:41.424456 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.424465 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.428953 4998 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Dec 03 16:06:41 crc kubenswrapper[4998]: E1203 16:06:41.476008 4998 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.64:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.546186 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.546255 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.546296 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.546385 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.546421 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.546447 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.546665 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.546711 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.648274 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.648371 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.648407 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.648473 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.648494 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.648553 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.648491 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.648556 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.648963 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.649165 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.649276 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.649312 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.649363 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.649385 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.649415 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.649460 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.689879 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50083a46-5fd2-4241-8a20-67ae6cd5f6b6" path="/var/lib/kubelet/pods/50083a46-5fd2-4241-8a20-67ae6cd5f6b6/volumes" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.690776 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56ad3380-4861-49d1-8758-7b1e27f74560" path="/var/lib/kubelet/pods/56ad3380-4861-49d1-8758-7b1e27f74560/volumes" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.691385 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97fb2ff7-4894-4858-99aa-1a54a3119d33" path="/var/lib/kubelet/pods/97fb2ff7-4894-4858-99aa-1a54a3119d33/volumes" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.692881 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d66602dc-ef95-4eec-989f-8e42dc4bcd02" path="/var/lib/kubelet/pods/d66602dc-ef95-4eec-989f-8e42dc4bcd02/volumes" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.750814 4998 generic.go:334] "Generic (PLEG): container finished" podID="20a499f0-e2b8-4204-9ff1-019ad9c78479" containerID="54c90d33a53ba3b54993ecc60273827e96597f341b9a19834c54af67e3a389da" exitCode=0 Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.750936 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"20a499f0-e2b8-4204-9ff1-019ad9c78479","Type":"ContainerDied","Data":"54c90d33a53ba3b54993ecc60273827e96597f341b9a19834c54af67e3a389da"} Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.751986 4998 status_manager.go:851] "Failed to get status for pod" podUID="20a499f0-e2b8-4204-9ff1-019ad9c78479" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.753039 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.754772 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.755823 4998 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56" exitCode=0 Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.755846 4998 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a" exitCode=0 Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.755854 4998 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2" exitCode=0 Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.755865 4998 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb" exitCode=2 Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.755915 4998 scope.go:117] "RemoveContainer" containerID="a9e50cfbeb6260a9d476f104812fbe03799c7cf3181ea201e07f31c9fc7f9955" Dec 03 16:06:41 crc kubenswrapper[4998]: I1203 16:06:41.777328 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:06:41 crc kubenswrapper[4998]: W1203 16:06:41.795201 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-58cc4986a1c96258ae3a0fe02c1f04c1af3c654dbc8caf5868910cc66b4dfc38 WatchSource:0}: Error finding container 58cc4986a1c96258ae3a0fe02c1f04c1af3c654dbc8caf5868910cc66b4dfc38: Status 404 returned error can't find the container with id 58cc4986a1c96258ae3a0fe02c1f04c1af3c654dbc8caf5868910cc66b4dfc38 Dec 03 16:06:41 crc kubenswrapper[4998]: E1203 16:06:41.798226 4998 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.64:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187dc03e109031a1 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 16:06:41.797493153 +0000 UTC m=+180.409193376,LastTimestamp:2025-12-03 16:06:41.797493153 +0000 UTC m=+180.409193376,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 16:06:42 crc kubenswrapper[4998]: I1203 16:06:42.772221 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 16:06:42 crc kubenswrapper[4998]: I1203 16:06:42.778210 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"f80441457b66da5573409afcf2058d53d519108bf2ac674809fa1ea9d3fcdca4"} Dec 03 16:06:42 crc kubenswrapper[4998]: I1203 16:06:42.778277 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"58cc4986a1c96258ae3a0fe02c1f04c1af3c654dbc8caf5868910cc66b4dfc38"} Dec 03 16:06:42 crc kubenswrapper[4998]: E1203 16:06:42.779239 4998 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.64:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:06:42 crc kubenswrapper[4998]: I1203 16:06:42.779318 4998 status_manager.go:851] "Failed to get status for pod" podUID="20a499f0-e2b8-4204-9ff1-019ad9c78479" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:42 crc kubenswrapper[4998]: E1203 16:06:42.829300 4998 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:42 crc kubenswrapper[4998]: E1203 16:06:42.830241 4998 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:42 crc kubenswrapper[4998]: E1203 16:06:42.830718 4998 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:42 crc kubenswrapper[4998]: E1203 16:06:42.831387 4998 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:42 crc kubenswrapper[4998]: E1203 16:06:42.831925 4998 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:42 crc kubenswrapper[4998]: I1203 16:06:42.831964 4998 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 03 16:06:42 crc kubenswrapper[4998]: E1203 16:06:42.832326 4998 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" interval="200ms" Dec 03 16:06:43 crc kubenswrapper[4998]: E1203 16:06:43.033038 4998 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" interval="400ms" Dec 03 16:06:43 crc kubenswrapper[4998]: I1203 16:06:43.098367 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:06:43 crc kubenswrapper[4998]: I1203 16:06:43.099177 4998 status_manager.go:851] "Failed to get status for pod" podUID="20a499f0-e2b8-4204-9ff1-019ad9c78479" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:43 crc kubenswrapper[4998]: I1203 16:06:43.169274 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/20a499f0-e2b8-4204-9ff1-019ad9c78479-kube-api-access\") pod \"20a499f0-e2b8-4204-9ff1-019ad9c78479\" (UID: \"20a499f0-e2b8-4204-9ff1-019ad9c78479\") " Dec 03 16:06:43 crc kubenswrapper[4998]: I1203 16:06:43.169369 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/20a499f0-e2b8-4204-9ff1-019ad9c78479-var-lock\") pod \"20a499f0-e2b8-4204-9ff1-019ad9c78479\" (UID: \"20a499f0-e2b8-4204-9ff1-019ad9c78479\") " Dec 03 16:06:43 crc kubenswrapper[4998]: I1203 16:06:43.169403 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/20a499f0-e2b8-4204-9ff1-019ad9c78479-kubelet-dir\") pod \"20a499f0-e2b8-4204-9ff1-019ad9c78479\" (UID: \"20a499f0-e2b8-4204-9ff1-019ad9c78479\") " Dec 03 16:06:43 crc kubenswrapper[4998]: I1203 16:06:43.169624 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/20a499f0-e2b8-4204-9ff1-019ad9c78479-var-lock" (OuterVolumeSpecName: "var-lock") pod "20a499f0-e2b8-4204-9ff1-019ad9c78479" (UID: "20a499f0-e2b8-4204-9ff1-019ad9c78479"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:06:43 crc kubenswrapper[4998]: I1203 16:06:43.169679 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/20a499f0-e2b8-4204-9ff1-019ad9c78479-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "20a499f0-e2b8-4204-9ff1-019ad9c78479" (UID: "20a499f0-e2b8-4204-9ff1-019ad9c78479"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:06:43 crc kubenswrapper[4998]: I1203 16:06:43.177558 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20a499f0-e2b8-4204-9ff1-019ad9c78479-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "20a499f0-e2b8-4204-9ff1-019ad9c78479" (UID: "20a499f0-e2b8-4204-9ff1-019ad9c78479"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:06:43 crc kubenswrapper[4998]: I1203 16:06:43.270825 4998 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/20a499f0-e2b8-4204-9ff1-019ad9c78479-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:43 crc kubenswrapper[4998]: I1203 16:06:43.270879 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/20a499f0-e2b8-4204-9ff1-019ad9c78479-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:43 crc kubenswrapper[4998]: I1203 16:06:43.270900 4998 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/20a499f0-e2b8-4204-9ff1-019ad9c78479-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:43 crc kubenswrapper[4998]: E1203 16:06:43.434105 4998 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" interval="800ms" Dec 03 16:06:43 crc kubenswrapper[4998]: I1203 16:06:43.785618 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"20a499f0-e2b8-4204-9ff1-019ad9c78479","Type":"ContainerDied","Data":"93e55e5129b09ebd864a5f2fda6959349d62a322faff567e32258d48c7ee16f2"} Dec 03 16:06:43 crc kubenswrapper[4998]: I1203 16:06:43.786014 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93e55e5129b09ebd864a5f2fda6959349d62a322faff567e32258d48c7ee16f2" Dec 03 16:06:43 crc kubenswrapper[4998]: I1203 16:06:43.785706 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 16:06:43 crc kubenswrapper[4998]: I1203 16:06:43.790650 4998 status_manager.go:851] "Failed to get status for pod" podUID="20a499f0-e2b8-4204-9ff1-019ad9c78479" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:44 crc kubenswrapper[4998]: E1203 16:06:44.235833 4998 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" interval="1.6s" Dec 03 16:06:45 crc kubenswrapper[4998]: I1203 16:06:45.805617 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 16:06:45 crc kubenswrapper[4998]: I1203 16:06:45.807005 4998 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079" exitCode=0 Dec 03 16:06:45 crc kubenswrapper[4998]: E1203 16:06:45.837992 4998 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" interval="3.2s" Dec 03 16:06:46 crc kubenswrapper[4998]: I1203 16:06:46.112232 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 16:06:46 crc kubenswrapper[4998]: I1203 16:06:46.113168 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:06:46 crc kubenswrapper[4998]: I1203 16:06:46.113530 4998 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:46 crc kubenswrapper[4998]: I1203 16:06:46.113916 4998 status_manager.go:851] "Failed to get status for pod" podUID="20a499f0-e2b8-4204-9ff1-019ad9c78479" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:46 crc kubenswrapper[4998]: I1203 16:06:46.214115 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 16:06:46 crc kubenswrapper[4998]: I1203 16:06:46.214299 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 16:06:46 crc kubenswrapper[4998]: I1203 16:06:46.214350 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 16:06:46 crc kubenswrapper[4998]: I1203 16:06:46.214348 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:06:46 crc kubenswrapper[4998]: I1203 16:06:46.214385 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:06:46 crc kubenswrapper[4998]: I1203 16:06:46.214482 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:06:46 crc kubenswrapper[4998]: I1203 16:06:46.214702 4998 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:46 crc kubenswrapper[4998]: I1203 16:06:46.214725 4998 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:46 crc kubenswrapper[4998]: I1203 16:06:46.214743 4998 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 16:06:46 crc kubenswrapper[4998]: I1203 16:06:46.817573 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 16:06:46 crc kubenswrapper[4998]: I1203 16:06:46.818902 4998 scope.go:117] "RemoveContainer" containerID="9169e33ef455a6a1e83fa50436261373ef2037a938c138bc82a9105d715aaa56" Dec 03 16:06:46 crc kubenswrapper[4998]: I1203 16:06:46.819147 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:06:46 crc kubenswrapper[4998]: I1203 16:06:46.844963 4998 scope.go:117] "RemoveContainer" containerID="2372bfbc61aeb81b46f2364693eaf6f3f7835970cf9a33dd0cd57ff39443cb8a" Dec 03 16:06:46 crc kubenswrapper[4998]: I1203 16:06:46.848587 4998 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:46 crc kubenswrapper[4998]: I1203 16:06:46.849320 4998 status_manager.go:851] "Failed to get status for pod" podUID="20a499f0-e2b8-4204-9ff1-019ad9c78479" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:46 crc kubenswrapper[4998]: I1203 16:06:46.871907 4998 scope.go:117] "RemoveContainer" containerID="ad87fa9e7529cedbc8133d4182114d62e27d43cb4f90877a8ade274286618cc2" Dec 03 16:06:46 crc kubenswrapper[4998]: I1203 16:06:46.889951 4998 scope.go:117] "RemoveContainer" containerID="8a94473557e78cb3c7238dee4046343941ab4c8dc15bd12c263e41aa175a79eb" Dec 03 16:06:46 crc kubenswrapper[4998]: I1203 16:06:46.909221 4998 scope.go:117] "RemoveContainer" containerID="08b8a69f349804eeacf751ad9257cbbe614fabeaaaa469b6834e908db93fa079" Dec 03 16:06:46 crc kubenswrapper[4998]: I1203 16:06:46.927954 4998 scope.go:117] "RemoveContainer" containerID="ae469694805fa6ed1aec547c9e23063af1d12fc72c673c6484f52743b72069f4" Dec 03 16:06:47 crc kubenswrapper[4998]: I1203 16:06:47.688461 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 03 16:06:48 crc kubenswrapper[4998]: E1203 16:06:48.132035 4998 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.64:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187dc03e109031a1 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 16:06:41.797493153 +0000 UTC m=+180.409193376,LastTimestamp:2025-12-03 16:06:41.797493153 +0000 UTC m=+180.409193376,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 16:06:49 crc kubenswrapper[4998]: I1203 16:06:49.004000 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 16:06:49 crc kubenswrapper[4998]: I1203 16:06:49.004601 4998 status_manager.go:851] "Failed to get status for pod" podUID="20a499f0-e2b8-4204-9ff1-019ad9c78479" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:49 crc kubenswrapper[4998]: I1203 16:06:49.004926 4998 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:49 crc kubenswrapper[4998]: E1203 16:06:49.039470 4998 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" interval="6.4s" Dec 03 16:06:51 crc kubenswrapper[4998]: I1203 16:06:51.681991 4998 status_manager.go:851] "Failed to get status for pod" podUID="20a499f0-e2b8-4204-9ff1-019ad9c78479" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:51 crc kubenswrapper[4998]: I1203 16:06:51.683027 4998 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:51 crc kubenswrapper[4998]: E1203 16:06:51.746805 4998 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.64:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" volumeName="registry-storage" Dec 03 16:06:54 crc kubenswrapper[4998]: I1203 16:06:54.677860 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:06:54 crc kubenswrapper[4998]: I1203 16:06:54.679831 4998 status_manager.go:851] "Failed to get status for pod" podUID="20a499f0-e2b8-4204-9ff1-019ad9c78479" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:54 crc kubenswrapper[4998]: I1203 16:06:54.681263 4998 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:54 crc kubenswrapper[4998]: I1203 16:06:54.702093 4998 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b6dace3c-e5d0-43ab-9022-1c597f24e231" Dec 03 16:06:54 crc kubenswrapper[4998]: I1203 16:06:54.702163 4998 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b6dace3c-e5d0-43ab-9022-1c597f24e231" Dec 03 16:06:54 crc kubenswrapper[4998]: E1203 16:06:54.702835 4998 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:06:54 crc kubenswrapper[4998]: I1203 16:06:54.703513 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:06:54 crc kubenswrapper[4998]: W1203 16:06:54.734519 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-b9b054538b2b4704d623f4a20f80389ea98a988eca20cec41a549e8bb365817e WatchSource:0}: Error finding container b9b054538b2b4704d623f4a20f80389ea98a988eca20cec41a549e8bb365817e: Status 404 returned error can't find the container with id b9b054538b2b4704d623f4a20f80389ea98a988eca20cec41a549e8bb365817e Dec 03 16:06:54 crc kubenswrapper[4998]: I1203 16:06:54.878329 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b9b054538b2b4704d623f4a20f80389ea98a988eca20cec41a549e8bb365817e"} Dec 03 16:06:54 crc kubenswrapper[4998]: I1203 16:06:54.881457 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 16:06:54 crc kubenswrapper[4998]: I1203 16:06:54.881536 4998 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a" exitCode=1 Dec 03 16:06:54 crc kubenswrapper[4998]: I1203 16:06:54.881581 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a"} Dec 03 16:06:54 crc kubenswrapper[4998]: I1203 16:06:54.882246 4998 scope.go:117] "RemoveContainer" containerID="6b56f7a1d1561a60110d38ffd1ad459d3d931aff5efdb2c5b0d8eb0fe84ed87a" Dec 03 16:06:54 crc kubenswrapper[4998]: I1203 16:06:54.883325 4998 status_manager.go:851] "Failed to get status for pod" podUID="20a499f0-e2b8-4204-9ff1-019ad9c78479" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:54 crc kubenswrapper[4998]: I1203 16:06:54.883847 4998 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:54 crc kubenswrapper[4998]: I1203 16:06:54.884306 4998 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:55 crc kubenswrapper[4998]: I1203 16:06:55.178857 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:06:55 crc kubenswrapper[4998]: E1203 16:06:55.441173 4998 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" interval="7s" Dec 03 16:06:55 crc kubenswrapper[4998]: I1203 16:06:55.506711 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:06:55 crc kubenswrapper[4998]: I1203 16:06:55.894190 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 16:06:55 crc kubenswrapper[4998]: I1203 16:06:55.894352 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"af89da6537ef4c0d7991a442c41a93851edd8f36f3a4637b62c13116eac310c4"} Dec 03 16:06:55 crc kubenswrapper[4998]: I1203 16:06:55.895541 4998 status_manager.go:851] "Failed to get status for pod" podUID="20a499f0-e2b8-4204-9ff1-019ad9c78479" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:55 crc kubenswrapper[4998]: I1203 16:06:55.896209 4998 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:55 crc kubenswrapper[4998]: I1203 16:06:55.896600 4998 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="674d21b1a419923c34af6de466f4e6a39ab07191eed680b11bc09ea2093bf753" exitCode=0 Dec 03 16:06:55 crc kubenswrapper[4998]: I1203 16:06:55.896652 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"674d21b1a419923c34af6de466f4e6a39ab07191eed680b11bc09ea2093bf753"} Dec 03 16:06:55 crc kubenswrapper[4998]: I1203 16:06:55.896798 4998 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:55 crc kubenswrapper[4998]: I1203 16:06:55.897144 4998 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b6dace3c-e5d0-43ab-9022-1c597f24e231" Dec 03 16:06:55 crc kubenswrapper[4998]: I1203 16:06:55.897189 4998 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b6dace3c-e5d0-43ab-9022-1c597f24e231" Dec 03 16:06:55 crc kubenswrapper[4998]: I1203 16:06:55.897573 4998 status_manager.go:851] "Failed to get status for pod" podUID="20a499f0-e2b8-4204-9ff1-019ad9c78479" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:55 crc kubenswrapper[4998]: E1203 16:06:55.897719 4998 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:06:55 crc kubenswrapper[4998]: I1203 16:06:55.898000 4998 status_manager.go:851] "Failed to get status for pod" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" pod="openshift-network-diagnostics/network-check-target-xd92c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/pods/network-check-target-xd92c\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:55 crc kubenswrapper[4998]: I1203 16:06:55.898519 4998 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 03 16:06:56 crc kubenswrapper[4998]: I1203 16:06:56.908028 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"10ec79159a18b381f2cd18fabe98c152b15949e2652b948f445a4c48182a512f"} Dec 03 16:06:56 crc kubenswrapper[4998]: I1203 16:06:56.908071 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7213bfbb57644add3db4c709994175fda03cd9a6c1191785b9adfc5ab2f4152c"} Dec 03 16:06:57 crc kubenswrapper[4998]: I1203 16:06:57.111402 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:06:57 crc kubenswrapper[4998]: I1203 16:06:57.111740 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:06:57 crc kubenswrapper[4998]: I1203 16:06:57.360692 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:06:57 crc kubenswrapper[4998]: I1203 16:06:57.364968 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:06:57 crc kubenswrapper[4998]: I1203 16:06:57.914785 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7a0a99677880a340cd9633174582999952f73eb53065c4699f6c3b5896391346"} Dec 03 16:06:57 crc kubenswrapper[4998]: I1203 16:06:57.915056 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:06:58 crc kubenswrapper[4998]: I1203 16:06:58.925634 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"cf33849a8751dc1916e79ed3519199ae64ffaa1208df5913a335f2b5bd31cdc6"} Dec 03 16:06:58 crc kubenswrapper[4998]: I1203 16:06:58.926037 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d28781ab91994c592a3ae74c59b1f64fada5abd0cffa342fc11913ba77031736"} Dec 03 16:06:58 crc kubenswrapper[4998]: I1203 16:06:58.926148 4998 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b6dace3c-e5d0-43ab-9022-1c597f24e231" Dec 03 16:06:58 crc kubenswrapper[4998]: I1203 16:06:58.926189 4998 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b6dace3c-e5d0-43ab-9022-1c597f24e231" Dec 03 16:06:59 crc kubenswrapper[4998]: I1203 16:06:59.704113 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:06:59 crc kubenswrapper[4998]: I1203 16:06:59.704185 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:06:59 crc kubenswrapper[4998]: I1203 16:06:59.712727 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:07:03 crc kubenswrapper[4998]: I1203 16:07:03.950205 4998 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:07:04 crc kubenswrapper[4998]: I1203 16:07:04.703614 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:07:04 crc kubenswrapper[4998]: I1203 16:07:04.711427 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:07:04 crc kubenswrapper[4998]: I1203 16:07:04.714727 4998 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="3441e57b-712a-4f51-ab28-4d5430b8b469" Dec 03 16:07:04 crc kubenswrapper[4998]: I1203 16:07:04.985715 4998 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b6dace3c-e5d0-43ab-9022-1c597f24e231" Dec 03 16:07:04 crc kubenswrapper[4998]: I1203 16:07:04.985789 4998 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b6dace3c-e5d0-43ab-9022-1c597f24e231" Dec 03 16:07:05 crc kubenswrapper[4998]: I1203 16:07:05.512818 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 16:07:05 crc kubenswrapper[4998]: I1203 16:07:05.990381 4998 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b6dace3c-e5d0-43ab-9022-1c597f24e231" Dec 03 16:07:05 crc kubenswrapper[4998]: I1203 16:07:05.990671 4998 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b6dace3c-e5d0-43ab-9022-1c597f24e231" Dec 03 16:07:11 crc kubenswrapper[4998]: I1203 16:07:11.709559 4998 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="3441e57b-712a-4f51-ab28-4d5430b8b469" Dec 03 16:07:12 crc kubenswrapper[4998]: I1203 16:07:12.847977 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 16:07:13 crc kubenswrapper[4998]: I1203 16:07:13.199444 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 16:07:13 crc kubenswrapper[4998]: I1203 16:07:13.673310 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 16:07:13 crc kubenswrapper[4998]: I1203 16:07:13.715383 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 16:07:14 crc kubenswrapper[4998]: I1203 16:07:14.276631 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 16:07:14 crc kubenswrapper[4998]: I1203 16:07:14.319835 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 16:07:14 crc kubenswrapper[4998]: I1203 16:07:14.649231 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 16:07:14 crc kubenswrapper[4998]: I1203 16:07:14.671279 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 16:07:14 crc kubenswrapper[4998]: I1203 16:07:14.886060 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 16:07:15 crc kubenswrapper[4998]: I1203 16:07:15.093208 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 16:07:15 crc kubenswrapper[4998]: I1203 16:07:15.303843 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 16:07:15 crc kubenswrapper[4998]: I1203 16:07:15.388196 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 16:07:15 crc kubenswrapper[4998]: I1203 16:07:15.480054 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 16:07:15 crc kubenswrapper[4998]: I1203 16:07:15.576915 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 16:07:15 crc kubenswrapper[4998]: I1203 16:07:15.645105 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 16:07:15 crc kubenswrapper[4998]: I1203 16:07:15.649648 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 16:07:15 crc kubenswrapper[4998]: I1203 16:07:15.675204 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 16:07:15 crc kubenswrapper[4998]: I1203 16:07:15.956017 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 16:07:15 crc kubenswrapper[4998]: I1203 16:07:15.999590 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 16:07:16 crc kubenswrapper[4998]: I1203 16:07:16.131595 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 16:07:16 crc kubenswrapper[4998]: I1203 16:07:16.223512 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 16:07:16 crc kubenswrapper[4998]: I1203 16:07:16.428799 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 16:07:16 crc kubenswrapper[4998]: I1203 16:07:16.444539 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 16:07:16 crc kubenswrapper[4998]: I1203 16:07:16.463200 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 16:07:16 crc kubenswrapper[4998]: I1203 16:07:16.560178 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 16:07:16 crc kubenswrapper[4998]: I1203 16:07:16.569390 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 16:07:16 crc kubenswrapper[4998]: I1203 16:07:16.569904 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 16:07:16 crc kubenswrapper[4998]: I1203 16:07:16.626100 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 16:07:16 crc kubenswrapper[4998]: I1203 16:07:16.631682 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 16:07:16 crc kubenswrapper[4998]: I1203 16:07:16.838140 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 16:07:16 crc kubenswrapper[4998]: I1203 16:07:16.865972 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 16:07:16 crc kubenswrapper[4998]: I1203 16:07:16.966263 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 16:07:17 crc kubenswrapper[4998]: I1203 16:07:17.033058 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 16:07:17 crc kubenswrapper[4998]: I1203 16:07:17.033568 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 16:07:17 crc kubenswrapper[4998]: I1203 16:07:17.038821 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 16:07:17 crc kubenswrapper[4998]: I1203 16:07:17.039600 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 16:07:17 crc kubenswrapper[4998]: I1203 16:07:17.051994 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 16:07:17 crc kubenswrapper[4998]: I1203 16:07:17.061498 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 16:07:17 crc kubenswrapper[4998]: I1203 16:07:17.242220 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 16:07:17 crc kubenswrapper[4998]: I1203 16:07:17.306791 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 16:07:17 crc kubenswrapper[4998]: I1203 16:07:17.331800 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 16:07:17 crc kubenswrapper[4998]: I1203 16:07:17.343880 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 16:07:17 crc kubenswrapper[4998]: I1203 16:07:17.455771 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 16:07:17 crc kubenswrapper[4998]: I1203 16:07:17.523450 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 16:07:17 crc kubenswrapper[4998]: I1203 16:07:17.552720 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 16:07:17 crc kubenswrapper[4998]: I1203 16:07:17.638887 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 16:07:17 crc kubenswrapper[4998]: I1203 16:07:17.780314 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 16:07:18 crc kubenswrapper[4998]: I1203 16:07:18.086936 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 16:07:18 crc kubenswrapper[4998]: I1203 16:07:18.113779 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 16:07:18 crc kubenswrapper[4998]: I1203 16:07:18.345196 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 16:07:18 crc kubenswrapper[4998]: I1203 16:07:18.404265 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 16:07:18 crc kubenswrapper[4998]: I1203 16:07:18.437734 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 16:07:18 crc kubenswrapper[4998]: I1203 16:07:18.494217 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 16:07:18 crc kubenswrapper[4998]: I1203 16:07:18.495658 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 16:07:18 crc kubenswrapper[4998]: I1203 16:07:18.561003 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 16:07:18 crc kubenswrapper[4998]: I1203 16:07:18.629193 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 16:07:18 crc kubenswrapper[4998]: I1203 16:07:18.629728 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 16:07:18 crc kubenswrapper[4998]: I1203 16:07:18.636179 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 16:07:18 crc kubenswrapper[4998]: I1203 16:07:18.647899 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 16:07:18 crc kubenswrapper[4998]: I1203 16:07:18.655941 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 16:07:18 crc kubenswrapper[4998]: I1203 16:07:18.729386 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 16:07:18 crc kubenswrapper[4998]: I1203 16:07:18.747287 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 16:07:18 crc kubenswrapper[4998]: I1203 16:07:18.808481 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 16:07:18 crc kubenswrapper[4998]: I1203 16:07:18.977419 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 16:07:18 crc kubenswrapper[4998]: I1203 16:07:18.980420 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 16:07:19 crc kubenswrapper[4998]: I1203 16:07:19.012163 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 16:07:19 crc kubenswrapper[4998]: I1203 16:07:19.031913 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 16:07:19 crc kubenswrapper[4998]: I1203 16:07:19.172577 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 16:07:19 crc kubenswrapper[4998]: I1203 16:07:19.418335 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 16:07:19 crc kubenswrapper[4998]: I1203 16:07:19.491309 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 16:07:19 crc kubenswrapper[4998]: I1203 16:07:19.505418 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 16:07:19 crc kubenswrapper[4998]: I1203 16:07:19.623641 4998 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 16:07:19 crc kubenswrapper[4998]: I1203 16:07:19.686608 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 16:07:19 crc kubenswrapper[4998]: I1203 16:07:19.756504 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 16:07:19 crc kubenswrapper[4998]: I1203 16:07:19.794371 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 16:07:19 crc kubenswrapper[4998]: I1203 16:07:19.809398 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 16:07:19 crc kubenswrapper[4998]: I1203 16:07:19.809926 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 16:07:19 crc kubenswrapper[4998]: I1203 16:07:19.947886 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 16:07:19 crc kubenswrapper[4998]: I1203 16:07:19.950158 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 16:07:19 crc kubenswrapper[4998]: I1203 16:07:19.979947 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 16:07:19 crc kubenswrapper[4998]: I1203 16:07:19.990436 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 16:07:20 crc kubenswrapper[4998]: I1203 16:07:20.006575 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 16:07:20 crc kubenswrapper[4998]: I1203 16:07:20.228092 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 16:07:20 crc kubenswrapper[4998]: I1203 16:07:20.257406 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 16:07:20 crc kubenswrapper[4998]: I1203 16:07:20.288117 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 16:07:20 crc kubenswrapper[4998]: I1203 16:07:20.306888 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 16:07:20 crc kubenswrapper[4998]: I1203 16:07:20.314012 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 16:07:20 crc kubenswrapper[4998]: I1203 16:07:20.319508 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 16:07:20 crc kubenswrapper[4998]: I1203 16:07:20.325214 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 16:07:20 crc kubenswrapper[4998]: I1203 16:07:20.401684 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 16:07:20 crc kubenswrapper[4998]: I1203 16:07:20.493399 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 16:07:20 crc kubenswrapper[4998]: I1203 16:07:20.606101 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 16:07:20 crc kubenswrapper[4998]: I1203 16:07:20.617928 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 16:07:20 crc kubenswrapper[4998]: I1203 16:07:20.653237 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 16:07:20 crc kubenswrapper[4998]: I1203 16:07:20.675987 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 16:07:20 crc kubenswrapper[4998]: I1203 16:07:20.786319 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 16:07:20 crc kubenswrapper[4998]: I1203 16:07:20.852815 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 16:07:20 crc kubenswrapper[4998]: I1203 16:07:20.880857 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 16:07:20 crc kubenswrapper[4998]: I1203 16:07:20.970111 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 16:07:21 crc kubenswrapper[4998]: I1203 16:07:21.085284 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 16:07:21 crc kubenswrapper[4998]: I1203 16:07:21.137644 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 16:07:21 crc kubenswrapper[4998]: I1203 16:07:21.145579 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 16:07:21 crc kubenswrapper[4998]: I1203 16:07:21.168509 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 16:07:21 crc kubenswrapper[4998]: I1203 16:07:21.185298 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 16:07:21 crc kubenswrapper[4998]: I1203 16:07:21.188025 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 16:07:21 crc kubenswrapper[4998]: I1203 16:07:21.214254 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 16:07:21 crc kubenswrapper[4998]: I1203 16:07:21.216666 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 16:07:21 crc kubenswrapper[4998]: I1203 16:07:21.295047 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 16:07:21 crc kubenswrapper[4998]: I1203 16:07:21.371668 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 16:07:21 crc kubenswrapper[4998]: I1203 16:07:21.490133 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 16:07:21 crc kubenswrapper[4998]: I1203 16:07:21.498626 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 16:07:21 crc kubenswrapper[4998]: I1203 16:07:21.734281 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 16:07:21 crc kubenswrapper[4998]: I1203 16:07:21.796554 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 16:07:21 crc kubenswrapper[4998]: I1203 16:07:21.860893 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 16:07:21 crc kubenswrapper[4998]: I1203 16:07:21.913647 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 16:07:21 crc kubenswrapper[4998]: I1203 16:07:21.992593 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 16:07:22 crc kubenswrapper[4998]: I1203 16:07:22.084274 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 16:07:22 crc kubenswrapper[4998]: I1203 16:07:22.125927 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 16:07:22 crc kubenswrapper[4998]: I1203 16:07:22.151796 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 16:07:22 crc kubenswrapper[4998]: I1203 16:07:22.250175 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 16:07:22 crc kubenswrapper[4998]: I1203 16:07:22.261897 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 16:07:22 crc kubenswrapper[4998]: I1203 16:07:22.267470 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 16:07:22 crc kubenswrapper[4998]: I1203 16:07:22.306427 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 16:07:22 crc kubenswrapper[4998]: I1203 16:07:22.341089 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 16:07:22 crc kubenswrapper[4998]: I1203 16:07:22.377458 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 16:07:22 crc kubenswrapper[4998]: I1203 16:07:22.415772 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 16:07:22 crc kubenswrapper[4998]: I1203 16:07:22.484284 4998 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 16:07:22 crc kubenswrapper[4998]: I1203 16:07:22.498195 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 16:07:22 crc kubenswrapper[4998]: I1203 16:07:22.547715 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 16:07:22 crc kubenswrapper[4998]: I1203 16:07:22.680599 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 16:07:22 crc kubenswrapper[4998]: I1203 16:07:22.682011 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 16:07:22 crc kubenswrapper[4998]: I1203 16:07:22.738536 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 16:07:22 crc kubenswrapper[4998]: I1203 16:07:22.758687 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 16:07:22 crc kubenswrapper[4998]: I1203 16:07:22.806139 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 16:07:22 crc kubenswrapper[4998]: I1203 16:07:22.832185 4998 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 16:07:22 crc kubenswrapper[4998]: I1203 16:07:22.945488 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 16:07:22 crc kubenswrapper[4998]: I1203 16:07:22.977968 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 16:07:23 crc kubenswrapper[4998]: I1203 16:07:23.049037 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 16:07:23 crc kubenswrapper[4998]: I1203 16:07:23.140206 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 16:07:23 crc kubenswrapper[4998]: I1203 16:07:23.154798 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 16:07:23 crc kubenswrapper[4998]: I1203 16:07:23.158950 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 16:07:23 crc kubenswrapper[4998]: I1203 16:07:23.229226 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 16:07:23 crc kubenswrapper[4998]: I1203 16:07:23.229347 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 16:07:23 crc kubenswrapper[4998]: I1203 16:07:23.241444 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 16:07:23 crc kubenswrapper[4998]: I1203 16:07:23.294088 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 16:07:23 crc kubenswrapper[4998]: I1203 16:07:23.349623 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 16:07:23 crc kubenswrapper[4998]: I1203 16:07:23.350436 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 16:07:23 crc kubenswrapper[4998]: I1203 16:07:23.358392 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 16:07:23 crc kubenswrapper[4998]: I1203 16:07:23.471535 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 16:07:23 crc kubenswrapper[4998]: I1203 16:07:23.519991 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 16:07:23 crc kubenswrapper[4998]: I1203 16:07:23.573409 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 16:07:23 crc kubenswrapper[4998]: I1203 16:07:23.611458 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 16:07:23 crc kubenswrapper[4998]: I1203 16:07:23.619806 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 16:07:23 crc kubenswrapper[4998]: I1203 16:07:23.725629 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 16:07:23 crc kubenswrapper[4998]: I1203 16:07:23.752615 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 16:07:23 crc kubenswrapper[4998]: I1203 16:07:23.843193 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 16:07:23 crc kubenswrapper[4998]: I1203 16:07:23.879004 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 16:07:23 crc kubenswrapper[4998]: I1203 16:07:23.881600 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 16:07:23 crc kubenswrapper[4998]: I1203 16:07:23.905566 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 16:07:23 crc kubenswrapper[4998]: I1203 16:07:23.913037 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 16:07:23 crc kubenswrapper[4998]: I1203 16:07:23.915640 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 16:07:24 crc kubenswrapper[4998]: I1203 16:07:24.004114 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 16:07:24 crc kubenswrapper[4998]: I1203 16:07:24.039930 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 16:07:24 crc kubenswrapper[4998]: I1203 16:07:24.043820 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 16:07:24 crc kubenswrapper[4998]: I1203 16:07:24.101951 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 16:07:24 crc kubenswrapper[4998]: I1203 16:07:24.332744 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 16:07:24 crc kubenswrapper[4998]: I1203 16:07:24.384311 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 16:07:24 crc kubenswrapper[4998]: I1203 16:07:24.396840 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 16:07:24 crc kubenswrapper[4998]: I1203 16:07:24.431082 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 16:07:24 crc kubenswrapper[4998]: I1203 16:07:24.610644 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 16:07:24 crc kubenswrapper[4998]: I1203 16:07:24.648242 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 16:07:24 crc kubenswrapper[4998]: I1203 16:07:24.758539 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 16:07:24 crc kubenswrapper[4998]: I1203 16:07:24.872093 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 16:07:25 crc kubenswrapper[4998]: I1203 16:07:25.040180 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 16:07:25 crc kubenswrapper[4998]: I1203 16:07:25.089301 4998 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 16:07:25 crc kubenswrapper[4998]: I1203 16:07:25.126267 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 16:07:25 crc kubenswrapper[4998]: I1203 16:07:25.165471 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 16:07:25 crc kubenswrapper[4998]: I1203 16:07:25.171823 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 16:07:25 crc kubenswrapper[4998]: I1203 16:07:25.191479 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 16:07:25 crc kubenswrapper[4998]: I1203 16:07:25.208111 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 16:07:25 crc kubenswrapper[4998]: I1203 16:07:25.257490 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 16:07:25 crc kubenswrapper[4998]: I1203 16:07:25.262347 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 16:07:25 crc kubenswrapper[4998]: I1203 16:07:25.303007 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 16:07:25 crc kubenswrapper[4998]: I1203 16:07:25.328010 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 16:07:25 crc kubenswrapper[4998]: I1203 16:07:25.342862 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 16:07:25 crc kubenswrapper[4998]: I1203 16:07:25.357256 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 16:07:25 crc kubenswrapper[4998]: I1203 16:07:25.376553 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 16:07:25 crc kubenswrapper[4998]: I1203 16:07:25.440564 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 16:07:25 crc kubenswrapper[4998]: I1203 16:07:25.471731 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 16:07:25 crc kubenswrapper[4998]: I1203 16:07:25.476088 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 16:07:25 crc kubenswrapper[4998]: I1203 16:07:25.543253 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 16:07:25 crc kubenswrapper[4998]: I1203 16:07:25.647437 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 16:07:25 crc kubenswrapper[4998]: I1203 16:07:25.736469 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 16:07:25 crc kubenswrapper[4998]: I1203 16:07:25.800971 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 16:07:25 crc kubenswrapper[4998]: I1203 16:07:25.945951 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 16:07:25 crc kubenswrapper[4998]: I1203 16:07:25.993149 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 16:07:26 crc kubenswrapper[4998]: I1203 16:07:26.017341 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 16:07:26 crc kubenswrapper[4998]: I1203 16:07:26.180908 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 16:07:26 crc kubenswrapper[4998]: I1203 16:07:26.274825 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 16:07:26 crc kubenswrapper[4998]: I1203 16:07:26.329643 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 16:07:26 crc kubenswrapper[4998]: I1203 16:07:26.337361 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 16:07:26 crc kubenswrapper[4998]: I1203 16:07:26.341060 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 16:07:26 crc kubenswrapper[4998]: I1203 16:07:26.341651 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 16:07:26 crc kubenswrapper[4998]: I1203 16:07:26.344368 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 16:07:26 crc kubenswrapper[4998]: I1203 16:07:26.636052 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 16:07:26 crc kubenswrapper[4998]: I1203 16:07:26.668852 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 16:07:26 crc kubenswrapper[4998]: I1203 16:07:26.676114 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 16:07:26 crc kubenswrapper[4998]: I1203 16:07:26.778234 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 16:07:26 crc kubenswrapper[4998]: I1203 16:07:26.823275 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 16:07:26 crc kubenswrapper[4998]: I1203 16:07:26.831423 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 16:07:26 crc kubenswrapper[4998]: I1203 16:07:26.834519 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.009229 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.027037 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.110908 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.110980 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.111040 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.111853 4998 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8"} pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.111944 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" containerID="cri-o://7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8" gracePeriod=600 Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.114310 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.168661 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.170170 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.214634 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.215653 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.373241 4998 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.383465 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.383561 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-546468998b-hqtjr"] Dec 03 16:07:27 crc kubenswrapper[4998]: E1203 16:07:27.383928 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20a499f0-e2b8-4204-9ff1-019ad9c78479" containerName="installer" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.383983 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="20a499f0-e2b8-4204-9ff1-019ad9c78479" containerName="installer" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.384209 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="20a499f0-e2b8-4204-9ff1-019ad9c78479" containerName="installer" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.384824 4998 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b6dace3c-e5d0-43ab-9022-1c597f24e231" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.385041 4998 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b6dace3c-e5d0-43ab-9022-1c597f24e231" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.385092 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.392042 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.393865 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.396452 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.396937 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.397229 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.397538 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.405175 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.405282 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.405360 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.405284 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.405466 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.405797 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.405838 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.412254 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.412966 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.423918 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.431201 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=24.431176866 podStartE2EDuration="24.431176866s" podCreationTimestamp="2025-12-03 16:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:07:27.421587442 +0000 UTC m=+226.033287665" watchObservedRunningTime="2025-12-03 16:07:27.431176866 +0000 UTC m=+226.042877099" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.434988 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.455271 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-system-session\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.455529 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcs6k\" (UniqueName: \"kubernetes.io/projected/62018d54-7199-41f9-a79a-c1891536d2a8-kube-api-access-kcs6k\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.455632 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.455726 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-system-router-certs\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.455953 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-user-template-error\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.456035 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-user-template-login\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.456113 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.456218 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/62018d54-7199-41f9-a79a-c1891536d2a8-audit-policies\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.456300 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.456401 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.456443 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.456517 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/62018d54-7199-41f9-a79a-c1891536d2a8-audit-dir\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.456581 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.456635 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-system-service-ca\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.557734 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-system-session\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.557808 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcs6k\" (UniqueName: \"kubernetes.io/projected/62018d54-7199-41f9-a79a-c1891536d2a8-kube-api-access-kcs6k\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.557837 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.558939 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-system-router-certs\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.558971 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-user-template-error\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.558993 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-user-template-login\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.559021 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.559050 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/62018d54-7199-41f9-a79a-c1891536d2a8-audit-policies\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.559072 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.559104 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.559124 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.559148 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/62018d54-7199-41f9-a79a-c1891536d2a8-audit-dir\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.559167 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.559185 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-system-service-ca\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.559898 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-system-service-ca\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.560636 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.565062 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-system-router-certs\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.565108 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/62018d54-7199-41f9-a79a-c1891536d2a8-audit-dir\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.568169 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.572043 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.573401 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/62018d54-7199-41f9-a79a-c1891536d2a8-audit-policies\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.573767 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.574626 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-system-session\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.575962 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.581577 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-user-template-error\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.581689 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-user-template-login\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.582021 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/62018d54-7199-41f9-a79a-c1891536d2a8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.582056 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcs6k\" (UniqueName: \"kubernetes.io/projected/62018d54-7199-41f9-a79a-c1891536d2a8-kube-api-access-kcs6k\") pod \"oauth-openshift-546468998b-hqtjr\" (UID: \"62018d54-7199-41f9-a79a-c1891536d2a8\") " pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.597852 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.609568 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.648792 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.717297 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:27 crc kubenswrapper[4998]: I1203 16:07:27.882202 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 16:07:28 crc kubenswrapper[4998]: I1203 16:07:28.018025 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 16:07:28 crc kubenswrapper[4998]: I1203 16:07:28.033405 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 16:07:28 crc kubenswrapper[4998]: I1203 16:07:28.132404 4998 generic.go:334] "Generic (PLEG): container finished" podID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerID="7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8" exitCode=0 Dec 03 16:07:28 crc kubenswrapper[4998]: I1203 16:07:28.132564 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerDied","Data":"7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8"} Dec 03 16:07:28 crc kubenswrapper[4998]: I1203 16:07:28.132674 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"4005a05bf5e77097685a266b8f7990957d3876e9377b8eac327681f73b327265"} Dec 03 16:07:28 crc kubenswrapper[4998]: W1203 16:07:28.183797 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62018d54_7199_41f9_a79a_c1891536d2a8.slice/crio-aee113447b6792c3f5b29db7814b5fa44b1907dc931e6a59149efcf9d372da9b WatchSource:0}: Error finding container aee113447b6792c3f5b29db7814b5fa44b1907dc931e6a59149efcf9d372da9b: Status 404 returned error can't find the container with id aee113447b6792c3f5b29db7814b5fa44b1907dc931e6a59149efcf9d372da9b Dec 03 16:07:28 crc kubenswrapper[4998]: I1203 16:07:28.187279 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-546468998b-hqtjr"] Dec 03 16:07:28 crc kubenswrapper[4998]: I1203 16:07:28.378979 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 16:07:28 crc kubenswrapper[4998]: I1203 16:07:28.393702 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 16:07:28 crc kubenswrapper[4998]: I1203 16:07:28.450467 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 16:07:28 crc kubenswrapper[4998]: I1203 16:07:28.553475 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 16:07:28 crc kubenswrapper[4998]: I1203 16:07:28.570516 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 16:07:28 crc kubenswrapper[4998]: I1203 16:07:28.591941 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 16:07:28 crc kubenswrapper[4998]: I1203 16:07:28.839278 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 16:07:28 crc kubenswrapper[4998]: I1203 16:07:28.900376 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 16:07:29 crc kubenswrapper[4998]: I1203 16:07:29.142033 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" event={"ID":"62018d54-7199-41f9-a79a-c1891536d2a8","Type":"ContainerStarted","Data":"301e69c142deb142649012f920f21bbeeb0b9462b5902546e70b99bc5a3c7f16"} Dec 03 16:07:29 crc kubenswrapper[4998]: I1203 16:07:29.142102 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" event={"ID":"62018d54-7199-41f9-a79a-c1891536d2a8","Type":"ContainerStarted","Data":"aee113447b6792c3f5b29db7814b5fa44b1907dc931e6a59149efcf9d372da9b"} Dec 03 16:07:29 crc kubenswrapper[4998]: I1203 16:07:29.168274 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" podStartSLOduration=76.168252772 podStartE2EDuration="1m16.168252772s" podCreationTimestamp="2025-12-03 16:06:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:07:29.168009066 +0000 UTC m=+227.779709379" watchObservedRunningTime="2025-12-03 16:07:29.168252772 +0000 UTC m=+227.779952995" Dec 03 16:07:29 crc kubenswrapper[4998]: I1203 16:07:29.193288 4998 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 16:07:29 crc kubenswrapper[4998]: I1203 16:07:29.378322 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 16:07:29 crc kubenswrapper[4998]: I1203 16:07:29.724337 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 16:07:30 crc kubenswrapper[4998]: I1203 16:07:30.147470 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:30 crc kubenswrapper[4998]: I1203 16:07:30.156027 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-546468998b-hqtjr" Dec 03 16:07:37 crc kubenswrapper[4998]: I1203 16:07:37.000219 4998 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 16:07:37 crc kubenswrapper[4998]: I1203 16:07:37.000926 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://f80441457b66da5573409afcf2058d53d519108bf2ac674809fa1ea9d3fcdca4" gracePeriod=5 Dec 03 16:07:42 crc kubenswrapper[4998]: I1203 16:07:42.221448 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 16:07:42 crc kubenswrapper[4998]: I1203 16:07:42.221916 4998 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="f80441457b66da5573409afcf2058d53d519108bf2ac674809fa1ea9d3fcdca4" exitCode=137 Dec 03 16:07:42 crc kubenswrapper[4998]: I1203 16:07:42.575259 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 16:07:42 crc kubenswrapper[4998]: I1203 16:07:42.575325 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:07:42 crc kubenswrapper[4998]: I1203 16:07:42.708117 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 16:07:42 crc kubenswrapper[4998]: I1203 16:07:42.708208 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 16:07:42 crc kubenswrapper[4998]: I1203 16:07:42.708241 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 16:07:42 crc kubenswrapper[4998]: I1203 16:07:42.708264 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:07:42 crc kubenswrapper[4998]: I1203 16:07:42.708289 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 16:07:42 crc kubenswrapper[4998]: I1203 16:07:42.708320 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 16:07:42 crc kubenswrapper[4998]: I1203 16:07:42.708396 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:07:42 crc kubenswrapper[4998]: I1203 16:07:42.708440 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:07:42 crc kubenswrapper[4998]: I1203 16:07:42.708486 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:07:42 crc kubenswrapper[4998]: I1203 16:07:42.708985 4998 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 16:07:42 crc kubenswrapper[4998]: I1203 16:07:42.709056 4998 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 16:07:42 crc kubenswrapper[4998]: I1203 16:07:42.709087 4998 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 03 16:07:42 crc kubenswrapper[4998]: I1203 16:07:42.709113 4998 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 03 16:07:42 crc kubenswrapper[4998]: I1203 16:07:42.719701 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:07:42 crc kubenswrapper[4998]: I1203 16:07:42.811155 4998 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 16:07:43 crc kubenswrapper[4998]: I1203 16:07:43.230639 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 16:07:43 crc kubenswrapper[4998]: I1203 16:07:43.230731 4998 scope.go:117] "RemoveContainer" containerID="f80441457b66da5573409afcf2058d53d519108bf2ac674809fa1ea9d3fcdca4" Dec 03 16:07:43 crc kubenswrapper[4998]: I1203 16:07:43.230803 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 16:07:43 crc kubenswrapper[4998]: I1203 16:07:43.684880 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.011578 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2kprn"] Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.012244 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" podUID="c2ff7dad-a4da-4243-ab1e-21b47230504d" containerName="controller-manager" containerID="cri-o://c7e0680cfd0e3670e44d7ed7f08e097b2c31d9a9ba81e3d06050a2cf6c2630bc" gracePeriod=30 Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.142202 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9"] Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.142615 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9" podUID="6024dcd8-96d6-4c42-9670-42e3608cd7ee" containerName="route-controller-manager" containerID="cri-o://3eee04750fd3bb67978aba152c91e073fc417746f33a9b9b563171b94743b1a2" gracePeriod=30 Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.250080 4998 generic.go:334] "Generic (PLEG): container finished" podID="c2ff7dad-a4da-4243-ab1e-21b47230504d" containerID="c7e0680cfd0e3670e44d7ed7f08e097b2c31d9a9ba81e3d06050a2cf6c2630bc" exitCode=0 Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.250290 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" event={"ID":"c2ff7dad-a4da-4243-ab1e-21b47230504d","Type":"ContainerDied","Data":"c7e0680cfd0e3670e44d7ed7f08e097b2c31d9a9ba81e3d06050a2cf6c2630bc"} Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.357475 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.446304 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2ff7dad-a4da-4243-ab1e-21b47230504d-serving-cert\") pod \"c2ff7dad-a4da-4243-ab1e-21b47230504d\" (UID: \"c2ff7dad-a4da-4243-ab1e-21b47230504d\") " Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.446364 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxdjj\" (UniqueName: \"kubernetes.io/projected/c2ff7dad-a4da-4243-ab1e-21b47230504d-kube-api-access-bxdjj\") pod \"c2ff7dad-a4da-4243-ab1e-21b47230504d\" (UID: \"c2ff7dad-a4da-4243-ab1e-21b47230504d\") " Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.446428 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c2ff7dad-a4da-4243-ab1e-21b47230504d-proxy-ca-bundles\") pod \"c2ff7dad-a4da-4243-ab1e-21b47230504d\" (UID: \"c2ff7dad-a4da-4243-ab1e-21b47230504d\") " Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.446484 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2ff7dad-a4da-4243-ab1e-21b47230504d-config\") pod \"c2ff7dad-a4da-4243-ab1e-21b47230504d\" (UID: \"c2ff7dad-a4da-4243-ab1e-21b47230504d\") " Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.446512 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c2ff7dad-a4da-4243-ab1e-21b47230504d-client-ca\") pod \"c2ff7dad-a4da-4243-ab1e-21b47230504d\" (UID: \"c2ff7dad-a4da-4243-ab1e-21b47230504d\") " Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.447266 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2ff7dad-a4da-4243-ab1e-21b47230504d-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "c2ff7dad-a4da-4243-ab1e-21b47230504d" (UID: "c2ff7dad-a4da-4243-ab1e-21b47230504d"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.447295 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2ff7dad-a4da-4243-ab1e-21b47230504d-client-ca" (OuterVolumeSpecName: "client-ca") pod "c2ff7dad-a4da-4243-ab1e-21b47230504d" (UID: "c2ff7dad-a4da-4243-ab1e-21b47230504d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.447386 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2ff7dad-a4da-4243-ab1e-21b47230504d-config" (OuterVolumeSpecName: "config") pod "c2ff7dad-a4da-4243-ab1e-21b47230504d" (UID: "c2ff7dad-a4da-4243-ab1e-21b47230504d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.454580 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2ff7dad-a4da-4243-ab1e-21b47230504d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c2ff7dad-a4da-4243-ab1e-21b47230504d" (UID: "c2ff7dad-a4da-4243-ab1e-21b47230504d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.454931 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2ff7dad-a4da-4243-ab1e-21b47230504d-kube-api-access-bxdjj" (OuterVolumeSpecName: "kube-api-access-bxdjj") pod "c2ff7dad-a4da-4243-ab1e-21b47230504d" (UID: "c2ff7dad-a4da-4243-ab1e-21b47230504d"). InnerVolumeSpecName "kube-api-access-bxdjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.473713 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9" Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.548166 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6024dcd8-96d6-4c42-9670-42e3608cd7ee-config\") pod \"6024dcd8-96d6-4c42-9670-42e3608cd7ee\" (UID: \"6024dcd8-96d6-4c42-9670-42e3608cd7ee\") " Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.548237 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6024dcd8-96d6-4c42-9670-42e3608cd7ee-client-ca\") pod \"6024dcd8-96d6-4c42-9670-42e3608cd7ee\" (UID: \"6024dcd8-96d6-4c42-9670-42e3608cd7ee\") " Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.548279 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6024dcd8-96d6-4c42-9670-42e3608cd7ee-serving-cert\") pod \"6024dcd8-96d6-4c42-9670-42e3608cd7ee\" (UID: \"6024dcd8-96d6-4c42-9670-42e3608cd7ee\") " Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.548325 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzn2p\" (UniqueName: \"kubernetes.io/projected/6024dcd8-96d6-4c42-9670-42e3608cd7ee-kube-api-access-tzn2p\") pod \"6024dcd8-96d6-4c42-9670-42e3608cd7ee\" (UID: \"6024dcd8-96d6-4c42-9670-42e3608cd7ee\") " Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.548619 4998 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c2ff7dad-a4da-4243-ab1e-21b47230504d-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.548638 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2ff7dad-a4da-4243-ab1e-21b47230504d-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.548649 4998 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c2ff7dad-a4da-4243-ab1e-21b47230504d-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.548663 4998 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2ff7dad-a4da-4243-ab1e-21b47230504d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.548673 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxdjj\" (UniqueName: \"kubernetes.io/projected/c2ff7dad-a4da-4243-ab1e-21b47230504d-kube-api-access-bxdjj\") on node \"crc\" DevicePath \"\"" Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.549029 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6024dcd8-96d6-4c42-9670-42e3608cd7ee-client-ca" (OuterVolumeSpecName: "client-ca") pod "6024dcd8-96d6-4c42-9670-42e3608cd7ee" (UID: "6024dcd8-96d6-4c42-9670-42e3608cd7ee"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.549298 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6024dcd8-96d6-4c42-9670-42e3608cd7ee-config" (OuterVolumeSpecName: "config") pod "6024dcd8-96d6-4c42-9670-42e3608cd7ee" (UID: "6024dcd8-96d6-4c42-9670-42e3608cd7ee"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.551577 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6024dcd8-96d6-4c42-9670-42e3608cd7ee-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6024dcd8-96d6-4c42-9670-42e3608cd7ee" (UID: "6024dcd8-96d6-4c42-9670-42e3608cd7ee"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.552444 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6024dcd8-96d6-4c42-9670-42e3608cd7ee-kube-api-access-tzn2p" (OuterVolumeSpecName: "kube-api-access-tzn2p") pod "6024dcd8-96d6-4c42-9670-42e3608cd7ee" (UID: "6024dcd8-96d6-4c42-9670-42e3608cd7ee"). InnerVolumeSpecName "kube-api-access-tzn2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.649853 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6024dcd8-96d6-4c42-9670-42e3608cd7ee-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.649891 4998 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6024dcd8-96d6-4c42-9670-42e3608cd7ee-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.649903 4998 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6024dcd8-96d6-4c42-9670-42e3608cd7ee-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:07:45 crc kubenswrapper[4998]: I1203 16:07:45.649916 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzn2p\" (UniqueName: \"kubernetes.io/projected/6024dcd8-96d6-4c42-9670-42e3608cd7ee-kube-api-access-tzn2p\") on node \"crc\" DevicePath \"\"" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.147316 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-869cd85dbd-vbpb6"] Dec 03 16:07:46 crc kubenswrapper[4998]: E1203 16:07:46.147612 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2ff7dad-a4da-4243-ab1e-21b47230504d" containerName="controller-manager" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.147631 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2ff7dad-a4da-4243-ab1e-21b47230504d" containerName="controller-manager" Dec 03 16:07:46 crc kubenswrapper[4998]: E1203 16:07:46.147661 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.147673 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 16:07:46 crc kubenswrapper[4998]: E1203 16:07:46.147694 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6024dcd8-96d6-4c42-9670-42e3608cd7ee" containerName="route-controller-manager" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.147709 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="6024dcd8-96d6-4c42-9670-42e3608cd7ee" containerName="route-controller-manager" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.147896 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2ff7dad-a4da-4243-ab1e-21b47230504d" containerName="controller-manager" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.147915 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="6024dcd8-96d6-4c42-9670-42e3608cd7ee" containerName="route-controller-manager" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.147938 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.148420 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-869cd85dbd-vbpb6" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.153504 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-869cd85dbd-vbpb6"] Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.256423 4998 generic.go:334] "Generic (PLEG): container finished" podID="6024dcd8-96d6-4c42-9670-42e3608cd7ee" containerID="3eee04750fd3bb67978aba152c91e073fc417746f33a9b9b563171b94743b1a2" exitCode=0 Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.256494 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.256489 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9" event={"ID":"6024dcd8-96d6-4c42-9670-42e3608cd7ee","Type":"ContainerDied","Data":"3eee04750fd3bb67978aba152c91e073fc417746f33a9b9b563171b94743b1a2"} Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.256865 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9" event={"ID":"6024dcd8-96d6-4c42-9670-42e3608cd7ee","Type":"ContainerDied","Data":"a1ef71f0e0b935d309143ccff1a8427f30cc84e052bf2b5d70fe371f7150b2b7"} Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.256892 4998 scope.go:117] "RemoveContainer" containerID="3eee04750fd3bb67978aba152c91e073fc417746f33a9b9b563171b94743b1a2" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.258411 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.258842 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb6wb\" (UniqueName: \"kubernetes.io/projected/e699eafe-310d-48f5-9165-5b2a68b953f6-kube-api-access-pb6wb\") pod \"controller-manager-869cd85dbd-vbpb6\" (UID: \"e699eafe-310d-48f5-9165-5b2a68b953f6\") " pod="openshift-controller-manager/controller-manager-869cd85dbd-vbpb6" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.258895 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e699eafe-310d-48f5-9165-5b2a68b953f6-client-ca\") pod \"controller-manager-869cd85dbd-vbpb6\" (UID: \"e699eafe-310d-48f5-9165-5b2a68b953f6\") " pod="openshift-controller-manager/controller-manager-869cd85dbd-vbpb6" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.258956 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e699eafe-310d-48f5-9165-5b2a68b953f6-proxy-ca-bundles\") pod \"controller-manager-869cd85dbd-vbpb6\" (UID: \"e699eafe-310d-48f5-9165-5b2a68b953f6\") " pod="openshift-controller-manager/controller-manager-869cd85dbd-vbpb6" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.259079 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2kprn" event={"ID":"c2ff7dad-a4da-4243-ab1e-21b47230504d","Type":"ContainerDied","Data":"1debfca7c9ac2dd15259c876cdeaf36a666d9b1e9969b08ad0388c33ae038c76"} Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.259097 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e699eafe-310d-48f5-9165-5b2a68b953f6-serving-cert\") pod \"controller-manager-869cd85dbd-vbpb6\" (UID: \"e699eafe-310d-48f5-9165-5b2a68b953f6\") " pod="openshift-controller-manager/controller-manager-869cd85dbd-vbpb6" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.259154 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e699eafe-310d-48f5-9165-5b2a68b953f6-config\") pod \"controller-manager-869cd85dbd-vbpb6\" (UID: \"e699eafe-310d-48f5-9165-5b2a68b953f6\") " pod="openshift-controller-manager/controller-manager-869cd85dbd-vbpb6" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.279947 4998 scope.go:117] "RemoveContainer" containerID="3eee04750fd3bb67978aba152c91e073fc417746f33a9b9b563171b94743b1a2" Dec 03 16:07:46 crc kubenswrapper[4998]: E1203 16:07:46.280443 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3eee04750fd3bb67978aba152c91e073fc417746f33a9b9b563171b94743b1a2\": container with ID starting with 3eee04750fd3bb67978aba152c91e073fc417746f33a9b9b563171b94743b1a2 not found: ID does not exist" containerID="3eee04750fd3bb67978aba152c91e073fc417746f33a9b9b563171b94743b1a2" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.280495 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eee04750fd3bb67978aba152c91e073fc417746f33a9b9b563171b94743b1a2"} err="failed to get container status \"3eee04750fd3bb67978aba152c91e073fc417746f33a9b9b563171b94743b1a2\": rpc error: code = NotFound desc = could not find container \"3eee04750fd3bb67978aba152c91e073fc417746f33a9b9b563171b94743b1a2\": container with ID starting with 3eee04750fd3bb67978aba152c91e073fc417746f33a9b9b563171b94743b1a2 not found: ID does not exist" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.280532 4998 scope.go:117] "RemoveContainer" containerID="c7e0680cfd0e3670e44d7ed7f08e097b2c31d9a9ba81e3d06050a2cf6c2630bc" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.287596 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9"] Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.287942 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-tvqq9"] Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.291099 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2kprn"] Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.294223 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2kprn"] Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.360566 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e699eafe-310d-48f5-9165-5b2a68b953f6-proxy-ca-bundles\") pod \"controller-manager-869cd85dbd-vbpb6\" (UID: \"e699eafe-310d-48f5-9165-5b2a68b953f6\") " pod="openshift-controller-manager/controller-manager-869cd85dbd-vbpb6" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.360634 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e699eafe-310d-48f5-9165-5b2a68b953f6-serving-cert\") pod \"controller-manager-869cd85dbd-vbpb6\" (UID: \"e699eafe-310d-48f5-9165-5b2a68b953f6\") " pod="openshift-controller-manager/controller-manager-869cd85dbd-vbpb6" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.360660 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e699eafe-310d-48f5-9165-5b2a68b953f6-config\") pod \"controller-manager-869cd85dbd-vbpb6\" (UID: \"e699eafe-310d-48f5-9165-5b2a68b953f6\") " pod="openshift-controller-manager/controller-manager-869cd85dbd-vbpb6" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.360691 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb6wb\" (UniqueName: \"kubernetes.io/projected/e699eafe-310d-48f5-9165-5b2a68b953f6-kube-api-access-pb6wb\") pod \"controller-manager-869cd85dbd-vbpb6\" (UID: \"e699eafe-310d-48f5-9165-5b2a68b953f6\") " pod="openshift-controller-manager/controller-manager-869cd85dbd-vbpb6" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.360734 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e699eafe-310d-48f5-9165-5b2a68b953f6-client-ca\") pod \"controller-manager-869cd85dbd-vbpb6\" (UID: \"e699eafe-310d-48f5-9165-5b2a68b953f6\") " pod="openshift-controller-manager/controller-manager-869cd85dbd-vbpb6" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.361701 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e699eafe-310d-48f5-9165-5b2a68b953f6-client-ca\") pod \"controller-manager-869cd85dbd-vbpb6\" (UID: \"e699eafe-310d-48f5-9165-5b2a68b953f6\") " pod="openshift-controller-manager/controller-manager-869cd85dbd-vbpb6" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.362766 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e699eafe-310d-48f5-9165-5b2a68b953f6-config\") pod \"controller-manager-869cd85dbd-vbpb6\" (UID: \"e699eafe-310d-48f5-9165-5b2a68b953f6\") " pod="openshift-controller-manager/controller-manager-869cd85dbd-vbpb6" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.364926 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e699eafe-310d-48f5-9165-5b2a68b953f6-proxy-ca-bundles\") pod \"controller-manager-869cd85dbd-vbpb6\" (UID: \"e699eafe-310d-48f5-9165-5b2a68b953f6\") " pod="openshift-controller-manager/controller-manager-869cd85dbd-vbpb6" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.365620 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e699eafe-310d-48f5-9165-5b2a68b953f6-serving-cert\") pod \"controller-manager-869cd85dbd-vbpb6\" (UID: \"e699eafe-310d-48f5-9165-5b2a68b953f6\") " pod="openshift-controller-manager/controller-manager-869cd85dbd-vbpb6" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.401602 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-869cd85dbd-vbpb6"] Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.401992 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb6wb\" (UniqueName: \"kubernetes.io/projected/e699eafe-310d-48f5-9165-5b2a68b953f6-kube-api-access-pb6wb\") pod \"controller-manager-869cd85dbd-vbpb6\" (UID: \"e699eafe-310d-48f5-9165-5b2a68b953f6\") " pod="openshift-controller-manager/controller-manager-869cd85dbd-vbpb6" Dec 03 16:07:46 crc kubenswrapper[4998]: E1203 16:07:46.402104 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-pb6wb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-controller-manager/controller-manager-869cd85dbd-vbpb6" podUID="e699eafe-310d-48f5-9165-5b2a68b953f6" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.445506 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz"] Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.446146 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.447686 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.447688 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.449141 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.449795 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.449808 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.451109 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.466727 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz"] Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.563642 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ae0bcea-dffa-471b-84eb-55298fbd29b5-config\") pod \"route-controller-manager-59877c47b5-j99tz\" (UID: \"1ae0bcea-dffa-471b-84eb-55298fbd29b5\") " pod="openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.563807 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1ae0bcea-dffa-471b-84eb-55298fbd29b5-client-ca\") pod \"route-controller-manager-59877c47b5-j99tz\" (UID: \"1ae0bcea-dffa-471b-84eb-55298fbd29b5\") " pod="openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.563881 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ae0bcea-dffa-471b-84eb-55298fbd29b5-serving-cert\") pod \"route-controller-manager-59877c47b5-j99tz\" (UID: \"1ae0bcea-dffa-471b-84eb-55298fbd29b5\") " pod="openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.563951 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhjgc\" (UniqueName: \"kubernetes.io/projected/1ae0bcea-dffa-471b-84eb-55298fbd29b5-kube-api-access-rhjgc\") pod \"route-controller-manager-59877c47b5-j99tz\" (UID: \"1ae0bcea-dffa-471b-84eb-55298fbd29b5\") " pod="openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.665623 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhjgc\" (UniqueName: \"kubernetes.io/projected/1ae0bcea-dffa-471b-84eb-55298fbd29b5-kube-api-access-rhjgc\") pod \"route-controller-manager-59877c47b5-j99tz\" (UID: \"1ae0bcea-dffa-471b-84eb-55298fbd29b5\") " pod="openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.666127 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ae0bcea-dffa-471b-84eb-55298fbd29b5-config\") pod \"route-controller-manager-59877c47b5-j99tz\" (UID: \"1ae0bcea-dffa-471b-84eb-55298fbd29b5\") " pod="openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.666380 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1ae0bcea-dffa-471b-84eb-55298fbd29b5-client-ca\") pod \"route-controller-manager-59877c47b5-j99tz\" (UID: \"1ae0bcea-dffa-471b-84eb-55298fbd29b5\") " pod="openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.666572 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ae0bcea-dffa-471b-84eb-55298fbd29b5-serving-cert\") pod \"route-controller-manager-59877c47b5-j99tz\" (UID: \"1ae0bcea-dffa-471b-84eb-55298fbd29b5\") " pod="openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.667717 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ae0bcea-dffa-471b-84eb-55298fbd29b5-config\") pod \"route-controller-manager-59877c47b5-j99tz\" (UID: \"1ae0bcea-dffa-471b-84eb-55298fbd29b5\") " pod="openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.668338 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1ae0bcea-dffa-471b-84eb-55298fbd29b5-client-ca\") pod \"route-controller-manager-59877c47b5-j99tz\" (UID: \"1ae0bcea-dffa-471b-84eb-55298fbd29b5\") " pod="openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.674691 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ae0bcea-dffa-471b-84eb-55298fbd29b5-serving-cert\") pod \"route-controller-manager-59877c47b5-j99tz\" (UID: \"1ae0bcea-dffa-471b-84eb-55298fbd29b5\") " pod="openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.691287 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhjgc\" (UniqueName: \"kubernetes.io/projected/1ae0bcea-dffa-471b-84eb-55298fbd29b5-kube-api-access-rhjgc\") pod \"route-controller-manager-59877c47b5-j99tz\" (UID: \"1ae0bcea-dffa-471b-84eb-55298fbd29b5\") " pod="openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz" Dec 03 16:07:46 crc kubenswrapper[4998]: I1203 16:07:46.759150 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz" Dec 03 16:07:47 crc kubenswrapper[4998]: I1203 16:07:47.203618 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz"] Dec 03 16:07:47 crc kubenswrapper[4998]: W1203 16:07:47.215124 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ae0bcea_dffa_471b_84eb_55298fbd29b5.slice/crio-839dd84a48b98fedb523c2a1e7ad25425781a276cfd3b7c750ba8fb2f292e6c3 WatchSource:0}: Error finding container 839dd84a48b98fedb523c2a1e7ad25425781a276cfd3b7c750ba8fb2f292e6c3: Status 404 returned error can't find the container with id 839dd84a48b98fedb523c2a1e7ad25425781a276cfd3b7c750ba8fb2f292e6c3 Dec 03 16:07:47 crc kubenswrapper[4998]: I1203 16:07:47.270587 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz" event={"ID":"1ae0bcea-dffa-471b-84eb-55298fbd29b5","Type":"ContainerStarted","Data":"839dd84a48b98fedb523c2a1e7ad25425781a276cfd3b7c750ba8fb2f292e6c3"} Dec 03 16:07:47 crc kubenswrapper[4998]: I1203 16:07:47.273868 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-869cd85dbd-vbpb6" Dec 03 16:07:47 crc kubenswrapper[4998]: I1203 16:07:47.306121 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-869cd85dbd-vbpb6" Dec 03 16:07:47 crc kubenswrapper[4998]: I1203 16:07:47.376272 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e699eafe-310d-48f5-9165-5b2a68b953f6-config\") pod \"e699eafe-310d-48f5-9165-5b2a68b953f6\" (UID: \"e699eafe-310d-48f5-9165-5b2a68b953f6\") " Dec 03 16:07:47 crc kubenswrapper[4998]: I1203 16:07:47.376589 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pb6wb\" (UniqueName: \"kubernetes.io/projected/e699eafe-310d-48f5-9165-5b2a68b953f6-kube-api-access-pb6wb\") pod \"e699eafe-310d-48f5-9165-5b2a68b953f6\" (UID: \"e699eafe-310d-48f5-9165-5b2a68b953f6\") " Dec 03 16:07:47 crc kubenswrapper[4998]: I1203 16:07:47.376815 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e699eafe-310d-48f5-9165-5b2a68b953f6-proxy-ca-bundles\") pod \"e699eafe-310d-48f5-9165-5b2a68b953f6\" (UID: \"e699eafe-310d-48f5-9165-5b2a68b953f6\") " Dec 03 16:07:47 crc kubenswrapper[4998]: I1203 16:07:47.376987 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e699eafe-310d-48f5-9165-5b2a68b953f6-serving-cert\") pod \"e699eafe-310d-48f5-9165-5b2a68b953f6\" (UID: \"e699eafe-310d-48f5-9165-5b2a68b953f6\") " Dec 03 16:07:47 crc kubenswrapper[4998]: I1203 16:07:47.377133 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e699eafe-310d-48f5-9165-5b2a68b953f6-client-ca\") pod \"e699eafe-310d-48f5-9165-5b2a68b953f6\" (UID: \"e699eafe-310d-48f5-9165-5b2a68b953f6\") " Dec 03 16:07:47 crc kubenswrapper[4998]: I1203 16:07:47.377957 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e699eafe-310d-48f5-9165-5b2a68b953f6-client-ca" (OuterVolumeSpecName: "client-ca") pod "e699eafe-310d-48f5-9165-5b2a68b953f6" (UID: "e699eafe-310d-48f5-9165-5b2a68b953f6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:07:47 crc kubenswrapper[4998]: I1203 16:07:47.378668 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e699eafe-310d-48f5-9165-5b2a68b953f6-config" (OuterVolumeSpecName: "config") pod "e699eafe-310d-48f5-9165-5b2a68b953f6" (UID: "e699eafe-310d-48f5-9165-5b2a68b953f6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:07:47 crc kubenswrapper[4998]: I1203 16:07:47.378882 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e699eafe-310d-48f5-9165-5b2a68b953f6-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e699eafe-310d-48f5-9165-5b2a68b953f6" (UID: "e699eafe-310d-48f5-9165-5b2a68b953f6"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:07:47 crc kubenswrapper[4998]: I1203 16:07:47.382127 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e699eafe-310d-48f5-9165-5b2a68b953f6-kube-api-access-pb6wb" (OuterVolumeSpecName: "kube-api-access-pb6wb") pod "e699eafe-310d-48f5-9165-5b2a68b953f6" (UID: "e699eafe-310d-48f5-9165-5b2a68b953f6"). InnerVolumeSpecName "kube-api-access-pb6wb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:07:47 crc kubenswrapper[4998]: I1203 16:07:47.385486 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e699eafe-310d-48f5-9165-5b2a68b953f6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e699eafe-310d-48f5-9165-5b2a68b953f6" (UID: "e699eafe-310d-48f5-9165-5b2a68b953f6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:07:47 crc kubenswrapper[4998]: I1203 16:07:47.479013 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e699eafe-310d-48f5-9165-5b2a68b953f6-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:07:47 crc kubenswrapper[4998]: I1203 16:07:47.479076 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pb6wb\" (UniqueName: \"kubernetes.io/projected/e699eafe-310d-48f5-9165-5b2a68b953f6-kube-api-access-pb6wb\") on node \"crc\" DevicePath \"\"" Dec 03 16:07:47 crc kubenswrapper[4998]: I1203 16:07:47.479105 4998 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e699eafe-310d-48f5-9165-5b2a68b953f6-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 16:07:47 crc kubenswrapper[4998]: I1203 16:07:47.479131 4998 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e699eafe-310d-48f5-9165-5b2a68b953f6-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:07:47 crc kubenswrapper[4998]: I1203 16:07:47.479155 4998 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e699eafe-310d-48f5-9165-5b2a68b953f6-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:07:47 crc kubenswrapper[4998]: I1203 16:07:47.688803 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6024dcd8-96d6-4c42-9670-42e3608cd7ee" path="/var/lib/kubelet/pods/6024dcd8-96d6-4c42-9670-42e3608cd7ee/volumes" Dec 03 16:07:47 crc kubenswrapper[4998]: I1203 16:07:47.690171 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2ff7dad-a4da-4243-ab1e-21b47230504d" path="/var/lib/kubelet/pods/c2ff7dad-a4da-4243-ab1e-21b47230504d/volumes" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.284456 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-869cd85dbd-vbpb6" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.284467 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz" event={"ID":"1ae0bcea-dffa-471b-84eb-55298fbd29b5","Type":"ContainerStarted","Data":"f8dbf1581fc65eacd6eb521d63a265d6647da761e10618c9362bb2f2bac0113d"} Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.285137 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.291502 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.316364 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz" podStartSLOduration=2.3163327479999998 podStartE2EDuration="2.316332748s" podCreationTimestamp="2025-12-03 16:07:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:07:48.310110593 +0000 UTC m=+246.921810856" watchObservedRunningTime="2025-12-03 16:07:48.316332748 +0000 UTC m=+246.928033031" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.352619 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-76b96d6657-nxrzt"] Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.353726 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.355962 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.358685 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.359064 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.359081 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.359397 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.360928 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-869cd85dbd-vbpb6"] Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.361636 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.370253 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.370594 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-869cd85dbd-vbpb6"] Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.376194 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-76b96d6657-nxrzt"] Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.494067 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bf726557-696e-4157-8e60-a9529b272ebc-client-ca\") pod \"controller-manager-76b96d6657-nxrzt\" (UID: \"bf726557-696e-4157-8e60-a9529b272ebc\") " pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.494144 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqcmj\" (UniqueName: \"kubernetes.io/projected/bf726557-696e-4157-8e60-a9529b272ebc-kube-api-access-kqcmj\") pod \"controller-manager-76b96d6657-nxrzt\" (UID: \"bf726557-696e-4157-8e60-a9529b272ebc\") " pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.494175 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bf726557-696e-4157-8e60-a9529b272ebc-proxy-ca-bundles\") pod \"controller-manager-76b96d6657-nxrzt\" (UID: \"bf726557-696e-4157-8e60-a9529b272ebc\") " pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.494201 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf726557-696e-4157-8e60-a9529b272ebc-config\") pod \"controller-manager-76b96d6657-nxrzt\" (UID: \"bf726557-696e-4157-8e60-a9529b272ebc\") " pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.494223 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf726557-696e-4157-8e60-a9529b272ebc-serving-cert\") pod \"controller-manager-76b96d6657-nxrzt\" (UID: \"bf726557-696e-4157-8e60-a9529b272ebc\") " pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.595956 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqcmj\" (UniqueName: \"kubernetes.io/projected/bf726557-696e-4157-8e60-a9529b272ebc-kube-api-access-kqcmj\") pod \"controller-manager-76b96d6657-nxrzt\" (UID: \"bf726557-696e-4157-8e60-a9529b272ebc\") " pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.596004 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bf726557-696e-4157-8e60-a9529b272ebc-proxy-ca-bundles\") pod \"controller-manager-76b96d6657-nxrzt\" (UID: \"bf726557-696e-4157-8e60-a9529b272ebc\") " pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.596041 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf726557-696e-4157-8e60-a9529b272ebc-config\") pod \"controller-manager-76b96d6657-nxrzt\" (UID: \"bf726557-696e-4157-8e60-a9529b272ebc\") " pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.596070 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf726557-696e-4157-8e60-a9529b272ebc-serving-cert\") pod \"controller-manager-76b96d6657-nxrzt\" (UID: \"bf726557-696e-4157-8e60-a9529b272ebc\") " pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.596152 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bf726557-696e-4157-8e60-a9529b272ebc-client-ca\") pod \"controller-manager-76b96d6657-nxrzt\" (UID: \"bf726557-696e-4157-8e60-a9529b272ebc\") " pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.597275 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bf726557-696e-4157-8e60-a9529b272ebc-proxy-ca-bundles\") pod \"controller-manager-76b96d6657-nxrzt\" (UID: \"bf726557-696e-4157-8e60-a9529b272ebc\") " pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.597347 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bf726557-696e-4157-8e60-a9529b272ebc-client-ca\") pod \"controller-manager-76b96d6657-nxrzt\" (UID: \"bf726557-696e-4157-8e60-a9529b272ebc\") " pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.597545 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf726557-696e-4157-8e60-a9529b272ebc-config\") pod \"controller-manager-76b96d6657-nxrzt\" (UID: \"bf726557-696e-4157-8e60-a9529b272ebc\") " pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.601388 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf726557-696e-4157-8e60-a9529b272ebc-serving-cert\") pod \"controller-manager-76b96d6657-nxrzt\" (UID: \"bf726557-696e-4157-8e60-a9529b272ebc\") " pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.624312 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqcmj\" (UniqueName: \"kubernetes.io/projected/bf726557-696e-4157-8e60-a9529b272ebc-kube-api-access-kqcmj\") pod \"controller-manager-76b96d6657-nxrzt\" (UID: \"bf726557-696e-4157-8e60-a9529b272ebc\") " pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.682529 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" Dec 03 16:07:48 crc kubenswrapper[4998]: I1203 16:07:48.889887 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-76b96d6657-nxrzt"] Dec 03 16:07:49 crc kubenswrapper[4998]: I1203 16:07:49.290516 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" event={"ID":"bf726557-696e-4157-8e60-a9529b272ebc","Type":"ContainerStarted","Data":"1367a387fc30210372420acba995659e9f0d13651320834168d547d8e41805b6"} Dec 03 16:07:49 crc kubenswrapper[4998]: I1203 16:07:49.291447 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" event={"ID":"bf726557-696e-4157-8e60-a9529b272ebc","Type":"ContainerStarted","Data":"e8e71799358442f6a6e8009b79cf76ae502379eb08afd9976b5520bde4ae30d9"} Dec 03 16:07:49 crc kubenswrapper[4998]: I1203 16:07:49.310313 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" podStartSLOduration=3.310292296 podStartE2EDuration="3.310292296s" podCreationTimestamp="2025-12-03 16:07:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:07:49.307059 +0000 UTC m=+247.918759233" watchObservedRunningTime="2025-12-03 16:07:49.310292296 +0000 UTC m=+247.921992519" Dec 03 16:07:49 crc kubenswrapper[4998]: I1203 16:07:49.686892 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e699eafe-310d-48f5-9165-5b2a68b953f6" path="/var/lib/kubelet/pods/e699eafe-310d-48f5-9165-5b2a68b953f6/volumes" Dec 03 16:07:50 crc kubenswrapper[4998]: I1203 16:07:50.300574 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" Dec 03 16:07:50 crc kubenswrapper[4998]: I1203 16:07:50.306846 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" Dec 03 16:08:05 crc kubenswrapper[4998]: I1203 16:08:05.468362 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-76b96d6657-nxrzt"] Dec 03 16:08:05 crc kubenswrapper[4998]: I1203 16:08:05.469272 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" podUID="bf726557-696e-4157-8e60-a9529b272ebc" containerName="controller-manager" containerID="cri-o://1367a387fc30210372420acba995659e9f0d13651320834168d547d8e41805b6" gracePeriod=30 Dec 03 16:08:06 crc kubenswrapper[4998]: I1203 16:08:06.031568 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" Dec 03 16:08:06 crc kubenswrapper[4998]: I1203 16:08:06.138095 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf726557-696e-4157-8e60-a9529b272ebc-config\") pod \"bf726557-696e-4157-8e60-a9529b272ebc\" (UID: \"bf726557-696e-4157-8e60-a9529b272ebc\") " Dec 03 16:08:06 crc kubenswrapper[4998]: I1203 16:08:06.138153 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bf726557-696e-4157-8e60-a9529b272ebc-proxy-ca-bundles\") pod \"bf726557-696e-4157-8e60-a9529b272ebc\" (UID: \"bf726557-696e-4157-8e60-a9529b272ebc\") " Dec 03 16:08:06 crc kubenswrapper[4998]: I1203 16:08:06.138183 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bf726557-696e-4157-8e60-a9529b272ebc-client-ca\") pod \"bf726557-696e-4157-8e60-a9529b272ebc\" (UID: \"bf726557-696e-4157-8e60-a9529b272ebc\") " Dec 03 16:08:06 crc kubenswrapper[4998]: I1203 16:08:06.138213 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqcmj\" (UniqueName: \"kubernetes.io/projected/bf726557-696e-4157-8e60-a9529b272ebc-kube-api-access-kqcmj\") pod \"bf726557-696e-4157-8e60-a9529b272ebc\" (UID: \"bf726557-696e-4157-8e60-a9529b272ebc\") " Dec 03 16:08:06 crc kubenswrapper[4998]: I1203 16:08:06.138269 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf726557-696e-4157-8e60-a9529b272ebc-serving-cert\") pod \"bf726557-696e-4157-8e60-a9529b272ebc\" (UID: \"bf726557-696e-4157-8e60-a9529b272ebc\") " Dec 03 16:08:06 crc kubenswrapper[4998]: I1203 16:08:06.139048 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf726557-696e-4157-8e60-a9529b272ebc-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "bf726557-696e-4157-8e60-a9529b272ebc" (UID: "bf726557-696e-4157-8e60-a9529b272ebc"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:08:06 crc kubenswrapper[4998]: I1203 16:08:06.139142 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf726557-696e-4157-8e60-a9529b272ebc-config" (OuterVolumeSpecName: "config") pod "bf726557-696e-4157-8e60-a9529b272ebc" (UID: "bf726557-696e-4157-8e60-a9529b272ebc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:08:06 crc kubenswrapper[4998]: I1203 16:08:06.139637 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf726557-696e-4157-8e60-a9529b272ebc-client-ca" (OuterVolumeSpecName: "client-ca") pod "bf726557-696e-4157-8e60-a9529b272ebc" (UID: "bf726557-696e-4157-8e60-a9529b272ebc"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:08:06 crc kubenswrapper[4998]: I1203 16:08:06.144182 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf726557-696e-4157-8e60-a9529b272ebc-kube-api-access-kqcmj" (OuterVolumeSpecName: "kube-api-access-kqcmj") pod "bf726557-696e-4157-8e60-a9529b272ebc" (UID: "bf726557-696e-4157-8e60-a9529b272ebc"). InnerVolumeSpecName "kube-api-access-kqcmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:08:06 crc kubenswrapper[4998]: I1203 16:08:06.144185 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf726557-696e-4157-8e60-a9529b272ebc-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bf726557-696e-4157-8e60-a9529b272ebc" (UID: "bf726557-696e-4157-8e60-a9529b272ebc"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:08:06 crc kubenswrapper[4998]: I1203 16:08:06.239794 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf726557-696e-4157-8e60-a9529b272ebc-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:06 crc kubenswrapper[4998]: I1203 16:08:06.239836 4998 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bf726557-696e-4157-8e60-a9529b272ebc-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:06 crc kubenswrapper[4998]: I1203 16:08:06.239846 4998 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bf726557-696e-4157-8e60-a9529b272ebc-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:06 crc kubenswrapper[4998]: I1203 16:08:06.239856 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqcmj\" (UniqueName: \"kubernetes.io/projected/bf726557-696e-4157-8e60-a9529b272ebc-kube-api-access-kqcmj\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:06 crc kubenswrapper[4998]: I1203 16:08:06.239865 4998 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf726557-696e-4157-8e60-a9529b272ebc-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:06 crc kubenswrapper[4998]: I1203 16:08:06.391698 4998 generic.go:334] "Generic (PLEG): container finished" podID="bf726557-696e-4157-8e60-a9529b272ebc" containerID="1367a387fc30210372420acba995659e9f0d13651320834168d547d8e41805b6" exitCode=0 Dec 03 16:08:06 crc kubenswrapper[4998]: I1203 16:08:06.392004 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" event={"ID":"bf726557-696e-4157-8e60-a9529b272ebc","Type":"ContainerDied","Data":"1367a387fc30210372420acba995659e9f0d13651320834168d547d8e41805b6"} Dec 03 16:08:06 crc kubenswrapper[4998]: I1203 16:08:06.392030 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" event={"ID":"bf726557-696e-4157-8e60-a9529b272ebc","Type":"ContainerDied","Data":"e8e71799358442f6a6e8009b79cf76ae502379eb08afd9976b5520bde4ae30d9"} Dec 03 16:08:06 crc kubenswrapper[4998]: I1203 16:08:06.392052 4998 scope.go:117] "RemoveContainer" containerID="1367a387fc30210372420acba995659e9f0d13651320834168d547d8e41805b6" Dec 03 16:08:06 crc kubenswrapper[4998]: I1203 16:08:06.392158 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76b96d6657-nxrzt" Dec 03 16:08:06 crc kubenswrapper[4998]: I1203 16:08:06.418592 4998 scope.go:117] "RemoveContainer" containerID="1367a387fc30210372420acba995659e9f0d13651320834168d547d8e41805b6" Dec 03 16:08:06 crc kubenswrapper[4998]: E1203 16:08:06.419997 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1367a387fc30210372420acba995659e9f0d13651320834168d547d8e41805b6\": container with ID starting with 1367a387fc30210372420acba995659e9f0d13651320834168d547d8e41805b6 not found: ID does not exist" containerID="1367a387fc30210372420acba995659e9f0d13651320834168d547d8e41805b6" Dec 03 16:08:06 crc kubenswrapper[4998]: I1203 16:08:06.420030 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1367a387fc30210372420acba995659e9f0d13651320834168d547d8e41805b6"} err="failed to get container status \"1367a387fc30210372420acba995659e9f0d13651320834168d547d8e41805b6\": rpc error: code = NotFound desc = could not find container \"1367a387fc30210372420acba995659e9f0d13651320834168d547d8e41805b6\": container with ID starting with 1367a387fc30210372420acba995659e9f0d13651320834168d547d8e41805b6 not found: ID does not exist" Dec 03 16:08:06 crc kubenswrapper[4998]: I1203 16:08:06.428102 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-76b96d6657-nxrzt"] Dec 03 16:08:06 crc kubenswrapper[4998]: I1203 16:08:06.434471 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-76b96d6657-nxrzt"] Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.162441 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7748d6f7fc-fqb8p"] Dec 03 16:08:07 crc kubenswrapper[4998]: E1203 16:08:07.162896 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf726557-696e-4157-8e60-a9529b272ebc" containerName="controller-manager" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.162925 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf726557-696e-4157-8e60-a9529b272ebc" containerName="controller-manager" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.163163 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf726557-696e-4157-8e60-a9529b272ebc" containerName="controller-manager" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.164102 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7748d6f7fc-fqb8p" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.166780 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.167517 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.168535 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.168544 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.169215 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.170139 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.172594 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7748d6f7fc-fqb8p"] Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.177401 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.253199 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9b895a7-7f64-4e02-88cb-1e5e4e65d523-proxy-ca-bundles\") pod \"controller-manager-7748d6f7fc-fqb8p\" (UID: \"a9b895a7-7f64-4e02-88cb-1e5e4e65d523\") " pod="openshift-controller-manager/controller-manager-7748d6f7fc-fqb8p" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.253284 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phzrr\" (UniqueName: \"kubernetes.io/projected/a9b895a7-7f64-4e02-88cb-1e5e4e65d523-kube-api-access-phzrr\") pod \"controller-manager-7748d6f7fc-fqb8p\" (UID: \"a9b895a7-7f64-4e02-88cb-1e5e4e65d523\") " pod="openshift-controller-manager/controller-manager-7748d6f7fc-fqb8p" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.253390 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9b895a7-7f64-4e02-88cb-1e5e4e65d523-config\") pod \"controller-manager-7748d6f7fc-fqb8p\" (UID: \"a9b895a7-7f64-4e02-88cb-1e5e4e65d523\") " pod="openshift-controller-manager/controller-manager-7748d6f7fc-fqb8p" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.253612 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a9b895a7-7f64-4e02-88cb-1e5e4e65d523-client-ca\") pod \"controller-manager-7748d6f7fc-fqb8p\" (UID: \"a9b895a7-7f64-4e02-88cb-1e5e4e65d523\") " pod="openshift-controller-manager/controller-manager-7748d6f7fc-fqb8p" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.253663 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9b895a7-7f64-4e02-88cb-1e5e4e65d523-serving-cert\") pod \"controller-manager-7748d6f7fc-fqb8p\" (UID: \"a9b895a7-7f64-4e02-88cb-1e5e4e65d523\") " pod="openshift-controller-manager/controller-manager-7748d6f7fc-fqb8p" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.355200 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9b895a7-7f64-4e02-88cb-1e5e4e65d523-proxy-ca-bundles\") pod \"controller-manager-7748d6f7fc-fqb8p\" (UID: \"a9b895a7-7f64-4e02-88cb-1e5e4e65d523\") " pod="openshift-controller-manager/controller-manager-7748d6f7fc-fqb8p" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.355289 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phzrr\" (UniqueName: \"kubernetes.io/projected/a9b895a7-7f64-4e02-88cb-1e5e4e65d523-kube-api-access-phzrr\") pod \"controller-manager-7748d6f7fc-fqb8p\" (UID: \"a9b895a7-7f64-4e02-88cb-1e5e4e65d523\") " pod="openshift-controller-manager/controller-manager-7748d6f7fc-fqb8p" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.355334 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9b895a7-7f64-4e02-88cb-1e5e4e65d523-config\") pod \"controller-manager-7748d6f7fc-fqb8p\" (UID: \"a9b895a7-7f64-4e02-88cb-1e5e4e65d523\") " pod="openshift-controller-manager/controller-manager-7748d6f7fc-fqb8p" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.355424 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a9b895a7-7f64-4e02-88cb-1e5e4e65d523-client-ca\") pod \"controller-manager-7748d6f7fc-fqb8p\" (UID: \"a9b895a7-7f64-4e02-88cb-1e5e4e65d523\") " pod="openshift-controller-manager/controller-manager-7748d6f7fc-fqb8p" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.355473 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9b895a7-7f64-4e02-88cb-1e5e4e65d523-serving-cert\") pod \"controller-manager-7748d6f7fc-fqb8p\" (UID: \"a9b895a7-7f64-4e02-88cb-1e5e4e65d523\") " pod="openshift-controller-manager/controller-manager-7748d6f7fc-fqb8p" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.356713 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a9b895a7-7f64-4e02-88cb-1e5e4e65d523-proxy-ca-bundles\") pod \"controller-manager-7748d6f7fc-fqb8p\" (UID: \"a9b895a7-7f64-4e02-88cb-1e5e4e65d523\") " pod="openshift-controller-manager/controller-manager-7748d6f7fc-fqb8p" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.358121 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a9b895a7-7f64-4e02-88cb-1e5e4e65d523-client-ca\") pod \"controller-manager-7748d6f7fc-fqb8p\" (UID: \"a9b895a7-7f64-4e02-88cb-1e5e4e65d523\") " pod="openshift-controller-manager/controller-manager-7748d6f7fc-fqb8p" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.358371 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9b895a7-7f64-4e02-88cb-1e5e4e65d523-config\") pod \"controller-manager-7748d6f7fc-fqb8p\" (UID: \"a9b895a7-7f64-4e02-88cb-1e5e4e65d523\") " pod="openshift-controller-manager/controller-manager-7748d6f7fc-fqb8p" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.361857 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9b895a7-7f64-4e02-88cb-1e5e4e65d523-serving-cert\") pod \"controller-manager-7748d6f7fc-fqb8p\" (UID: \"a9b895a7-7f64-4e02-88cb-1e5e4e65d523\") " pod="openshift-controller-manager/controller-manager-7748d6f7fc-fqb8p" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.377622 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phzrr\" (UniqueName: \"kubernetes.io/projected/a9b895a7-7f64-4e02-88cb-1e5e4e65d523-kube-api-access-phzrr\") pod \"controller-manager-7748d6f7fc-fqb8p\" (UID: \"a9b895a7-7f64-4e02-88cb-1e5e4e65d523\") " pod="openshift-controller-manager/controller-manager-7748d6f7fc-fqb8p" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.487243 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7748d6f7fc-fqb8p" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.691609 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf726557-696e-4157-8e60-a9529b272ebc" path="/var/lib/kubelet/pods/bf726557-696e-4157-8e60-a9529b272ebc/volumes" Dec 03 16:08:07 crc kubenswrapper[4998]: I1203 16:08:07.748421 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7748d6f7fc-fqb8p"] Dec 03 16:08:07 crc kubenswrapper[4998]: W1203 16:08:07.753785 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9b895a7_7f64_4e02_88cb_1e5e4e65d523.slice/crio-e79346e205b1a1853e108fa6b7d1b93d3a74c9064af8c7e6d47f1776bb36e740 WatchSource:0}: Error finding container e79346e205b1a1853e108fa6b7d1b93d3a74c9064af8c7e6d47f1776bb36e740: Status 404 returned error can't find the container with id e79346e205b1a1853e108fa6b7d1b93d3a74c9064af8c7e6d47f1776bb36e740 Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.042650 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-zq6nl"] Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.044089 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.062883 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-zq6nl"] Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.164326 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/67517c40-0016-42f2-8ab5-0ccc2db55801-bound-sa-token\") pod \"image-registry-66df7c8f76-zq6nl\" (UID: \"67517c40-0016-42f2-8ab5-0ccc2db55801\") " pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.164375 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/67517c40-0016-42f2-8ab5-0ccc2db55801-installation-pull-secrets\") pod \"image-registry-66df7c8f76-zq6nl\" (UID: \"67517c40-0016-42f2-8ab5-0ccc2db55801\") " pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.164404 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/67517c40-0016-42f2-8ab5-0ccc2db55801-registry-certificates\") pod \"image-registry-66df7c8f76-zq6nl\" (UID: \"67517c40-0016-42f2-8ab5-0ccc2db55801\") " pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.164568 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/67517c40-0016-42f2-8ab5-0ccc2db55801-trusted-ca\") pod \"image-registry-66df7c8f76-zq6nl\" (UID: \"67517c40-0016-42f2-8ab5-0ccc2db55801\") " pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.164712 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/67517c40-0016-42f2-8ab5-0ccc2db55801-ca-trust-extracted\") pod \"image-registry-66df7c8f76-zq6nl\" (UID: \"67517c40-0016-42f2-8ab5-0ccc2db55801\") " pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.164745 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/67517c40-0016-42f2-8ab5-0ccc2db55801-registry-tls\") pod \"image-registry-66df7c8f76-zq6nl\" (UID: \"67517c40-0016-42f2-8ab5-0ccc2db55801\") " pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.164794 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-zq6nl\" (UID: \"67517c40-0016-42f2-8ab5-0ccc2db55801\") " pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.164828 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prj8c\" (UniqueName: \"kubernetes.io/projected/67517c40-0016-42f2-8ab5-0ccc2db55801-kube-api-access-prj8c\") pod \"image-registry-66df7c8f76-zq6nl\" (UID: \"67517c40-0016-42f2-8ab5-0ccc2db55801\") " pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.185191 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-zq6nl\" (UID: \"67517c40-0016-42f2-8ab5-0ccc2db55801\") " pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.265923 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/67517c40-0016-42f2-8ab5-0ccc2db55801-ca-trust-extracted\") pod \"image-registry-66df7c8f76-zq6nl\" (UID: \"67517c40-0016-42f2-8ab5-0ccc2db55801\") " pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.265988 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/67517c40-0016-42f2-8ab5-0ccc2db55801-registry-tls\") pod \"image-registry-66df7c8f76-zq6nl\" (UID: \"67517c40-0016-42f2-8ab5-0ccc2db55801\") " pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.266021 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prj8c\" (UniqueName: \"kubernetes.io/projected/67517c40-0016-42f2-8ab5-0ccc2db55801-kube-api-access-prj8c\") pod \"image-registry-66df7c8f76-zq6nl\" (UID: \"67517c40-0016-42f2-8ab5-0ccc2db55801\") " pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.266057 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/67517c40-0016-42f2-8ab5-0ccc2db55801-bound-sa-token\") pod \"image-registry-66df7c8f76-zq6nl\" (UID: \"67517c40-0016-42f2-8ab5-0ccc2db55801\") " pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.266091 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/67517c40-0016-42f2-8ab5-0ccc2db55801-installation-pull-secrets\") pod \"image-registry-66df7c8f76-zq6nl\" (UID: \"67517c40-0016-42f2-8ab5-0ccc2db55801\") " pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.266126 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/67517c40-0016-42f2-8ab5-0ccc2db55801-registry-certificates\") pod \"image-registry-66df7c8f76-zq6nl\" (UID: \"67517c40-0016-42f2-8ab5-0ccc2db55801\") " pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.266171 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/67517c40-0016-42f2-8ab5-0ccc2db55801-trusted-ca\") pod \"image-registry-66df7c8f76-zq6nl\" (UID: \"67517c40-0016-42f2-8ab5-0ccc2db55801\") " pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.266504 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/67517c40-0016-42f2-8ab5-0ccc2db55801-ca-trust-extracted\") pod \"image-registry-66df7c8f76-zq6nl\" (UID: \"67517c40-0016-42f2-8ab5-0ccc2db55801\") " pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.267816 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/67517c40-0016-42f2-8ab5-0ccc2db55801-trusted-ca\") pod \"image-registry-66df7c8f76-zq6nl\" (UID: \"67517c40-0016-42f2-8ab5-0ccc2db55801\") " pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.268152 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/67517c40-0016-42f2-8ab5-0ccc2db55801-registry-certificates\") pod \"image-registry-66df7c8f76-zq6nl\" (UID: \"67517c40-0016-42f2-8ab5-0ccc2db55801\") " pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.273109 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/67517c40-0016-42f2-8ab5-0ccc2db55801-installation-pull-secrets\") pod \"image-registry-66df7c8f76-zq6nl\" (UID: \"67517c40-0016-42f2-8ab5-0ccc2db55801\") " pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.273172 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/67517c40-0016-42f2-8ab5-0ccc2db55801-registry-tls\") pod \"image-registry-66df7c8f76-zq6nl\" (UID: \"67517c40-0016-42f2-8ab5-0ccc2db55801\") " pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.283249 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prj8c\" (UniqueName: \"kubernetes.io/projected/67517c40-0016-42f2-8ab5-0ccc2db55801-kube-api-access-prj8c\") pod \"image-registry-66df7c8f76-zq6nl\" (UID: \"67517c40-0016-42f2-8ab5-0ccc2db55801\") " pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.297163 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/67517c40-0016-42f2-8ab5-0ccc2db55801-bound-sa-token\") pod \"image-registry-66df7c8f76-zq6nl\" (UID: \"67517c40-0016-42f2-8ab5-0ccc2db55801\") " pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.361936 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.416559 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7748d6f7fc-fqb8p" event={"ID":"a9b895a7-7f64-4e02-88cb-1e5e4e65d523","Type":"ContainerStarted","Data":"e79346e205b1a1853e108fa6b7d1b93d3a74c9064af8c7e6d47f1776bb36e740"} Dec 03 16:08:08 crc kubenswrapper[4998]: I1203 16:08:08.814074 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-zq6nl"] Dec 03 16:08:09 crc kubenswrapper[4998]: I1203 16:08:09.423357 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7748d6f7fc-fqb8p" event={"ID":"a9b895a7-7f64-4e02-88cb-1e5e4e65d523","Type":"ContainerStarted","Data":"4a2c2b00bb7dc4af363fff27f4c56d39f32e760e04bc223f3073fc7eb32f01d0"} Dec 03 16:08:09 crc kubenswrapper[4998]: I1203 16:08:09.424381 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7748d6f7fc-fqb8p" Dec 03 16:08:09 crc kubenswrapper[4998]: I1203 16:08:09.425533 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" event={"ID":"67517c40-0016-42f2-8ab5-0ccc2db55801","Type":"ContainerStarted","Data":"317e45aa3f8f0c8474236b4ad1f77d8cedd9b90b700b546a2318e0e2897ccab5"} Dec 03 16:08:09 crc kubenswrapper[4998]: I1203 16:08:09.425570 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" event={"ID":"67517c40-0016-42f2-8ab5-0ccc2db55801","Type":"ContainerStarted","Data":"37d04601276869f86598cf5a5f3bd039cbc8101523676af76149575b6a2cc8a4"} Dec 03 16:08:09 crc kubenswrapper[4998]: I1203 16:08:09.425940 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:09 crc kubenswrapper[4998]: I1203 16:08:09.431922 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7748d6f7fc-fqb8p" Dec 03 16:08:09 crc kubenswrapper[4998]: I1203 16:08:09.457106 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7748d6f7fc-fqb8p" podStartSLOduration=4.457082258 podStartE2EDuration="4.457082258s" podCreationTimestamp="2025-12-03 16:08:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:08:09.453160873 +0000 UTC m=+268.064861096" watchObservedRunningTime="2025-12-03 16:08:09.457082258 +0000 UTC m=+268.068782481" Dec 03 16:08:14 crc kubenswrapper[4998]: I1203 16:08:14.676586 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" podStartSLOduration=6.676568233 podStartE2EDuration="6.676568233s" podCreationTimestamp="2025-12-03 16:08:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:08:09.518066283 +0000 UTC m=+268.129766506" watchObservedRunningTime="2025-12-03 16:08:14.676568233 +0000 UTC m=+273.288268456" Dec 03 16:08:14 crc kubenswrapper[4998]: I1203 16:08:14.678743 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-84rhg"] Dec 03 16:08:14 crc kubenswrapper[4998]: I1203 16:08:14.679071 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-84rhg" podUID="8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea" containerName="registry-server" containerID="cri-o://08bd17c836a424fd644b4fbfe560085cf2d06aa7c9151a7a2bf0e1d29a30d34a" gracePeriod=30 Dec 03 16:08:14 crc kubenswrapper[4998]: I1203 16:08:14.696255 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rrdtb"] Dec 03 16:08:14 crc kubenswrapper[4998]: I1203 16:08:14.696640 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rrdtb" podUID="5ef027be-08d9-4731-bd35-34f7bb2fe43a" containerName="registry-server" containerID="cri-o://b4d411c20e1346b15e05f170e9d66fd9f1a449db05f1b05820f9b8dc0834578c" gracePeriod=30 Dec 03 16:08:14 crc kubenswrapper[4998]: I1203 16:08:14.707132 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fnxt7"] Dec 03 16:08:14 crc kubenswrapper[4998]: I1203 16:08:14.707358 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-fnxt7" podUID="8c715a96-664c-4b26-8d5a-523a0fb4a17c" containerName="marketplace-operator" containerID="cri-o://0865731cfb451c650cd07d5dcb67765501d216049957bb9aa334a71a6fd665e8" gracePeriod=30 Dec 03 16:08:14 crc kubenswrapper[4998]: I1203 16:08:14.717493 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dsh8q"] Dec 03 16:08:14 crc kubenswrapper[4998]: I1203 16:08:14.717806 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dsh8q" podUID="85eaaffb-675d-4283-9ef9-4765097f8a51" containerName="registry-server" containerID="cri-o://d3a607b5ca6a8a42d2ef8df7393dc1263c11c7009612744e258b02823a97806f" gracePeriod=30 Dec 03 16:08:14 crc kubenswrapper[4998]: I1203 16:08:14.728596 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6rx7t"] Dec 03 16:08:14 crc kubenswrapper[4998]: I1203 16:08:14.736047 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6rx7t" podUID="ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34" containerName="registry-server" containerID="cri-o://66b64f36635097dde5bf4df271ee7d55124f89c8c5c229fef9a8b1e3312645e8" gracePeriod=30 Dec 03 16:08:14 crc kubenswrapper[4998]: I1203 16:08:14.740513 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ts48f"] Dec 03 16:08:14 crc kubenswrapper[4998]: I1203 16:08:14.743441 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ts48f" Dec 03 16:08:14 crc kubenswrapper[4998]: I1203 16:08:14.765736 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ts48f"] Dec 03 16:08:14 crc kubenswrapper[4998]: I1203 16:08:14.853078 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4d7e0e87-75f9-437a-916a-f702c53d370b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ts48f\" (UID: \"4d7e0e87-75f9-437a-916a-f702c53d370b\") " pod="openshift-marketplace/marketplace-operator-79b997595-ts48f" Dec 03 16:08:14 crc kubenswrapper[4998]: I1203 16:08:14.853355 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4fw4\" (UniqueName: \"kubernetes.io/projected/4d7e0e87-75f9-437a-916a-f702c53d370b-kube-api-access-h4fw4\") pod \"marketplace-operator-79b997595-ts48f\" (UID: \"4d7e0e87-75f9-437a-916a-f702c53d370b\") " pod="openshift-marketplace/marketplace-operator-79b997595-ts48f" Dec 03 16:08:14 crc kubenswrapper[4998]: I1203 16:08:14.853466 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4d7e0e87-75f9-437a-916a-f702c53d370b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ts48f\" (UID: \"4d7e0e87-75f9-437a-916a-f702c53d370b\") " pod="openshift-marketplace/marketplace-operator-79b997595-ts48f" Dec 03 16:08:14 crc kubenswrapper[4998]: I1203 16:08:14.954588 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4d7e0e87-75f9-437a-916a-f702c53d370b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ts48f\" (UID: \"4d7e0e87-75f9-437a-916a-f702c53d370b\") " pod="openshift-marketplace/marketplace-operator-79b997595-ts48f" Dec 03 16:08:14 crc kubenswrapper[4998]: I1203 16:08:14.954981 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4fw4\" (UniqueName: \"kubernetes.io/projected/4d7e0e87-75f9-437a-916a-f702c53d370b-kube-api-access-h4fw4\") pod \"marketplace-operator-79b997595-ts48f\" (UID: \"4d7e0e87-75f9-437a-916a-f702c53d370b\") " pod="openshift-marketplace/marketplace-operator-79b997595-ts48f" Dec 03 16:08:14 crc kubenswrapper[4998]: I1203 16:08:14.955042 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4d7e0e87-75f9-437a-916a-f702c53d370b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ts48f\" (UID: \"4d7e0e87-75f9-437a-916a-f702c53d370b\") " pod="openshift-marketplace/marketplace-operator-79b997595-ts48f" Dec 03 16:08:14 crc kubenswrapper[4998]: I1203 16:08:14.957162 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4d7e0e87-75f9-437a-916a-f702c53d370b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ts48f\" (UID: \"4d7e0e87-75f9-437a-916a-f702c53d370b\") " pod="openshift-marketplace/marketplace-operator-79b997595-ts48f" Dec 03 16:08:14 crc kubenswrapper[4998]: I1203 16:08:14.973501 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4d7e0e87-75f9-437a-916a-f702c53d370b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ts48f\" (UID: \"4d7e0e87-75f9-437a-916a-f702c53d370b\") " pod="openshift-marketplace/marketplace-operator-79b997595-ts48f" Dec 03 16:08:14 crc kubenswrapper[4998]: I1203 16:08:14.978067 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4fw4\" (UniqueName: \"kubernetes.io/projected/4d7e0e87-75f9-437a-916a-f702c53d370b-kube-api-access-h4fw4\") pod \"marketplace-operator-79b997595-ts48f\" (UID: \"4d7e0e87-75f9-437a-916a-f702c53d370b\") " pod="openshift-marketplace/marketplace-operator-79b997595-ts48f" Dec 03 16:08:15 crc kubenswrapper[4998]: I1203 16:08:15.075284 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ts48f" Dec 03 16:08:15 crc kubenswrapper[4998]: I1203 16:08:15.459266 4998 generic.go:334] "Generic (PLEG): container finished" podID="8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea" containerID="08bd17c836a424fd644b4fbfe560085cf2d06aa7c9151a7a2bf0e1d29a30d34a" exitCode=0 Dec 03 16:08:15 crc kubenswrapper[4998]: I1203 16:08:15.459509 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84rhg" event={"ID":"8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea","Type":"ContainerDied","Data":"08bd17c836a424fd644b4fbfe560085cf2d06aa7c9151a7a2bf0e1d29a30d34a"} Dec 03 16:08:15 crc kubenswrapper[4998]: I1203 16:08:15.462827 4998 generic.go:334] "Generic (PLEG): container finished" podID="ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34" containerID="66b64f36635097dde5bf4df271ee7d55124f89c8c5c229fef9a8b1e3312645e8" exitCode=0 Dec 03 16:08:15 crc kubenswrapper[4998]: I1203 16:08:15.462871 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6rx7t" event={"ID":"ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34","Type":"ContainerDied","Data":"66b64f36635097dde5bf4df271ee7d55124f89c8c5c229fef9a8b1e3312645e8"} Dec 03 16:08:15 crc kubenswrapper[4998]: I1203 16:08:15.466988 4998 generic.go:334] "Generic (PLEG): container finished" podID="85eaaffb-675d-4283-9ef9-4765097f8a51" containerID="d3a607b5ca6a8a42d2ef8df7393dc1263c11c7009612744e258b02823a97806f" exitCode=0 Dec 03 16:08:15 crc kubenswrapper[4998]: I1203 16:08:15.467032 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dsh8q" event={"ID":"85eaaffb-675d-4283-9ef9-4765097f8a51","Type":"ContainerDied","Data":"d3a607b5ca6a8a42d2ef8df7393dc1263c11c7009612744e258b02823a97806f"} Dec 03 16:08:15 crc kubenswrapper[4998]: I1203 16:08:15.468267 4998 generic.go:334] "Generic (PLEG): container finished" podID="8c715a96-664c-4b26-8d5a-523a0fb4a17c" containerID="0865731cfb451c650cd07d5dcb67765501d216049957bb9aa334a71a6fd665e8" exitCode=0 Dec 03 16:08:15 crc kubenswrapper[4998]: I1203 16:08:15.468330 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fnxt7" event={"ID":"8c715a96-664c-4b26-8d5a-523a0fb4a17c","Type":"ContainerDied","Data":"0865731cfb451c650cd07d5dcb67765501d216049957bb9aa334a71a6fd665e8"} Dec 03 16:08:15 crc kubenswrapper[4998]: I1203 16:08:15.487913 4998 generic.go:334] "Generic (PLEG): container finished" podID="5ef027be-08d9-4731-bd35-34f7bb2fe43a" containerID="b4d411c20e1346b15e05f170e9d66fd9f1a449db05f1b05820f9b8dc0834578c" exitCode=0 Dec 03 16:08:15 crc kubenswrapper[4998]: I1203 16:08:15.487958 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rrdtb" event={"ID":"5ef027be-08d9-4731-bd35-34f7bb2fe43a","Type":"ContainerDied","Data":"b4d411c20e1346b15e05f170e9d66fd9f1a449db05f1b05820f9b8dc0834578c"} Dec 03 16:08:15 crc kubenswrapper[4998]: I1203 16:08:15.534523 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ts48f"] Dec 03 16:08:15 crc kubenswrapper[4998]: I1203 16:08:15.745712 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-fnxt7" Dec 03 16:08:15 crc kubenswrapper[4998]: I1203 16:08:15.865351 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c715a96-664c-4b26-8d5a-523a0fb4a17c-marketplace-trusted-ca\") pod \"8c715a96-664c-4b26-8d5a-523a0fb4a17c\" (UID: \"8c715a96-664c-4b26-8d5a-523a0fb4a17c\") " Dec 03 16:08:15 crc kubenswrapper[4998]: I1203 16:08:15.865399 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gt8xm\" (UniqueName: \"kubernetes.io/projected/8c715a96-664c-4b26-8d5a-523a0fb4a17c-kube-api-access-gt8xm\") pod \"8c715a96-664c-4b26-8d5a-523a0fb4a17c\" (UID: \"8c715a96-664c-4b26-8d5a-523a0fb4a17c\") " Dec 03 16:08:15 crc kubenswrapper[4998]: I1203 16:08:15.865492 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8c715a96-664c-4b26-8d5a-523a0fb4a17c-marketplace-operator-metrics\") pod \"8c715a96-664c-4b26-8d5a-523a0fb4a17c\" (UID: \"8c715a96-664c-4b26-8d5a-523a0fb4a17c\") " Dec 03 16:08:15 crc kubenswrapper[4998]: I1203 16:08:15.866277 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c715a96-664c-4b26-8d5a-523a0fb4a17c-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "8c715a96-664c-4b26-8d5a-523a0fb4a17c" (UID: "8c715a96-664c-4b26-8d5a-523a0fb4a17c"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:08:15 crc kubenswrapper[4998]: I1203 16:08:15.873830 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c715a96-664c-4b26-8d5a-523a0fb4a17c-kube-api-access-gt8xm" (OuterVolumeSpecName: "kube-api-access-gt8xm") pod "8c715a96-664c-4b26-8d5a-523a0fb4a17c" (UID: "8c715a96-664c-4b26-8d5a-523a0fb4a17c"). InnerVolumeSpecName "kube-api-access-gt8xm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:08:15 crc kubenswrapper[4998]: I1203 16:08:15.884829 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c715a96-664c-4b26-8d5a-523a0fb4a17c-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "8c715a96-664c-4b26-8d5a-523a0fb4a17c" (UID: "8c715a96-664c-4b26-8d5a-523a0fb4a17c"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:08:15 crc kubenswrapper[4998]: I1203 16:08:15.966601 4998 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8c715a96-664c-4b26-8d5a-523a0fb4a17c-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:15 crc kubenswrapper[4998]: I1203 16:08:15.966651 4998 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c715a96-664c-4b26-8d5a-523a0fb4a17c-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:15 crc kubenswrapper[4998]: I1203 16:08:15.966661 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gt8xm\" (UniqueName: \"kubernetes.io/projected/8c715a96-664c-4b26-8d5a-523a0fb4a17c-kube-api-access-gt8xm\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.054274 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dsh8q" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.085104 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rrdtb" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.086721 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6rx7t" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.113305 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-84rhg" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.168595 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkhch\" (UniqueName: \"kubernetes.io/projected/5ef027be-08d9-4731-bd35-34f7bb2fe43a-kube-api-access-dkhch\") pod \"5ef027be-08d9-4731-bd35-34f7bb2fe43a\" (UID: \"5ef027be-08d9-4731-bd35-34f7bb2fe43a\") " Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.168644 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ef027be-08d9-4731-bd35-34f7bb2fe43a-utilities\") pod \"5ef027be-08d9-4731-bd35-34f7bb2fe43a\" (UID: \"5ef027be-08d9-4731-bd35-34f7bb2fe43a\") " Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.168697 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ef027be-08d9-4731-bd35-34f7bb2fe43a-catalog-content\") pod \"5ef027be-08d9-4731-bd35-34f7bb2fe43a\" (UID: \"5ef027be-08d9-4731-bd35-34f7bb2fe43a\") " Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.168800 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34-utilities\") pod \"ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34\" (UID: \"ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34\") " Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.168818 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xk5vk\" (UniqueName: \"kubernetes.io/projected/ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34-kube-api-access-xk5vk\") pod \"ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34\" (UID: \"ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34\") " Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.168852 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85eaaffb-675d-4283-9ef9-4765097f8a51-utilities\") pod \"85eaaffb-675d-4283-9ef9-4765097f8a51\" (UID: \"85eaaffb-675d-4283-9ef9-4765097f8a51\") " Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.168923 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwf58\" (UniqueName: \"kubernetes.io/projected/85eaaffb-675d-4283-9ef9-4765097f8a51-kube-api-access-mwf58\") pod \"85eaaffb-675d-4283-9ef9-4765097f8a51\" (UID: \"85eaaffb-675d-4283-9ef9-4765097f8a51\") " Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.168943 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85eaaffb-675d-4283-9ef9-4765097f8a51-catalog-content\") pod \"85eaaffb-675d-4283-9ef9-4765097f8a51\" (UID: \"85eaaffb-675d-4283-9ef9-4765097f8a51\") " Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.168967 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34-catalog-content\") pod \"ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34\" (UID: \"ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34\") " Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.169550 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ef027be-08d9-4731-bd35-34f7bb2fe43a-utilities" (OuterVolumeSpecName: "utilities") pod "5ef027be-08d9-4731-bd35-34f7bb2fe43a" (UID: "5ef027be-08d9-4731-bd35-34f7bb2fe43a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.169948 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34-utilities" (OuterVolumeSpecName: "utilities") pod "ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34" (UID: "ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.173220 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85eaaffb-675d-4283-9ef9-4765097f8a51-kube-api-access-mwf58" (OuterVolumeSpecName: "kube-api-access-mwf58") pod "85eaaffb-675d-4283-9ef9-4765097f8a51" (UID: "85eaaffb-675d-4283-9ef9-4765097f8a51"). InnerVolumeSpecName "kube-api-access-mwf58". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.173519 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85eaaffb-675d-4283-9ef9-4765097f8a51-utilities" (OuterVolumeSpecName: "utilities") pod "85eaaffb-675d-4283-9ef9-4765097f8a51" (UID: "85eaaffb-675d-4283-9ef9-4765097f8a51"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.175192 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ef027be-08d9-4731-bd35-34f7bb2fe43a-kube-api-access-dkhch" (OuterVolumeSpecName: "kube-api-access-dkhch") pod "5ef027be-08d9-4731-bd35-34f7bb2fe43a" (UID: "5ef027be-08d9-4731-bd35-34f7bb2fe43a"). InnerVolumeSpecName "kube-api-access-dkhch". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.176264 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34-kube-api-access-xk5vk" (OuterVolumeSpecName: "kube-api-access-xk5vk") pod "ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34" (UID: "ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34"). InnerVolumeSpecName "kube-api-access-xk5vk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.208041 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85eaaffb-675d-4283-9ef9-4765097f8a51-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "85eaaffb-675d-4283-9ef9-4765097f8a51" (UID: "85eaaffb-675d-4283-9ef9-4765097f8a51"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.225379 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ef027be-08d9-4731-bd35-34f7bb2fe43a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ef027be-08d9-4731-bd35-34f7bb2fe43a" (UID: "5ef027be-08d9-4731-bd35-34f7bb2fe43a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.270517 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea-utilities\") pod \"8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea\" (UID: \"8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea\") " Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.270604 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea-catalog-content\") pod \"8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea\" (UID: \"8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea\") " Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.270700 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52sp8\" (UniqueName: \"kubernetes.io/projected/8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea-kube-api-access-52sp8\") pod \"8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea\" (UID: \"8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea\") " Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.270979 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwf58\" (UniqueName: \"kubernetes.io/projected/85eaaffb-675d-4283-9ef9-4765097f8a51-kube-api-access-mwf58\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.270991 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85eaaffb-675d-4283-9ef9-4765097f8a51-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.271000 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkhch\" (UniqueName: \"kubernetes.io/projected/5ef027be-08d9-4731-bd35-34f7bb2fe43a-kube-api-access-dkhch\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.271009 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ef027be-08d9-4731-bd35-34f7bb2fe43a-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.271017 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ef027be-08d9-4731-bd35-34f7bb2fe43a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.271025 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.271033 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xk5vk\" (UniqueName: \"kubernetes.io/projected/ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34-kube-api-access-xk5vk\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.271041 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85eaaffb-675d-4283-9ef9-4765097f8a51-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.272106 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea-utilities" (OuterVolumeSpecName: "utilities") pod "8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea" (UID: "8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.273847 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea-kube-api-access-52sp8" (OuterVolumeSpecName: "kube-api-access-52sp8") pod "8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea" (UID: "8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea"). InnerVolumeSpecName "kube-api-access-52sp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.293247 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34" (UID: "ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.319598 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea" (UID: "8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.372510 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.372550 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.372563 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52sp8\" (UniqueName: \"kubernetes.io/projected/8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea-kube-api-access-52sp8\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.372573 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.496598 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6rx7t" event={"ID":"ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34","Type":"ContainerDied","Data":"bc8c7fa6598914723871453ecb9bb50e56ab5d0b3b5c8b45291b26715320066d"} Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.496722 4998 scope.go:117] "RemoveContainer" containerID="66b64f36635097dde5bf4df271ee7d55124f89c8c5c229fef9a8b1e3312645e8" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.496602 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6rx7t" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.499123 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dsh8q" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.499120 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dsh8q" event={"ID":"85eaaffb-675d-4283-9ef9-4765097f8a51","Type":"ContainerDied","Data":"75950a1ada30242cdd65fa71d3945ab82a42c81a3da99dcec7e844fbd569e98c"} Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.501058 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fnxt7" event={"ID":"8c715a96-664c-4b26-8d5a-523a0fb4a17c","Type":"ContainerDied","Data":"bf682da7e63db18e5efb3089e9cf4cba51cc142a82e7c8e5581fb1b95efe6c4e"} Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.501108 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-fnxt7" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.504324 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rrdtb" event={"ID":"5ef027be-08d9-4731-bd35-34f7bb2fe43a","Type":"ContainerDied","Data":"19bbba31d6c74a4d6a8681fb879b07e78290a2dd9ff873ca370c05447adc0050"} Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.509788 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-ts48f" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.509930 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ts48f" event={"ID":"4d7e0e87-75f9-437a-916a-f702c53d370b","Type":"ContainerStarted","Data":"9d7f4eb1fa69a9552ade7772be4156bf68429b6e8c182f38b6aaa08a1a40c7ee"} Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.510001 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ts48f" event={"ID":"4d7e0e87-75f9-437a-916a-f702c53d370b","Type":"ContainerStarted","Data":"7a934c43b1530a90b4d60a7c9da0d49ea507ba0e5f103c9d31e3c534618b4b40"} Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.504563 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rrdtb" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.511161 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-84rhg" event={"ID":"8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea","Type":"ContainerDied","Data":"e17423948c3a9e9cc421a30ee51b7690d2edd560fea445f6ae4760ec1423a981"} Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.514223 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-84rhg" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.514315 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-ts48f" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.527600 4998 scope.go:117] "RemoveContainer" containerID="0c936724e87a152b1a722995b064278def6da26f462f12385824f59fef1feacf" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.540130 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-ts48f" podStartSLOduration=2.540103528 podStartE2EDuration="2.540103528s" podCreationTimestamp="2025-12-03 16:08:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:08:16.538496315 +0000 UTC m=+275.150196558" watchObservedRunningTime="2025-12-03 16:08:16.540103528 +0000 UTC m=+275.151803751" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.571985 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6rx7t"] Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.573080 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6rx7t"] Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.578410 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fnxt7"] Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.585915 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fnxt7"] Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.588149 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rrdtb"] Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.591504 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rrdtb"] Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.594578 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dsh8q"] Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.594623 4998 scope.go:117] "RemoveContainer" containerID="87f30a6956f51cf5e9b948c3d0533656461cabf031841ad2aaa6e49aa32d9dbc" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.601154 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dsh8q"] Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.610364 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-84rhg"] Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.613249 4998 scope.go:117] "RemoveContainer" containerID="d3a607b5ca6a8a42d2ef8df7393dc1263c11c7009612744e258b02823a97806f" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.614191 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-84rhg"] Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.628525 4998 scope.go:117] "RemoveContainer" containerID="87c4e338a88366f2254f40595c16d184e8d890b446fb058042ca84b832367e78" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.640895 4998 scope.go:117] "RemoveContainer" containerID="b13ec9ddb03a2d461bd4464b9402eed32b800bdd1e3d4fe4cebf0ec379f69ede" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.654729 4998 scope.go:117] "RemoveContainer" containerID="0865731cfb451c650cd07d5dcb67765501d216049957bb9aa334a71a6fd665e8" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.668027 4998 scope.go:117] "RemoveContainer" containerID="b4d411c20e1346b15e05f170e9d66fd9f1a449db05f1b05820f9b8dc0834578c" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.685933 4998 scope.go:117] "RemoveContainer" containerID="17b15688ad76e3b67d4cdbe192aa2e3129305e78031fb82916e6295b9ad77ffa" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.701929 4998 scope.go:117] "RemoveContainer" containerID="0b499c108a7efc0fae07e669d4d60cf49c27eb8e5f18280f27cf13f90a4f0f49" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.721051 4998 scope.go:117] "RemoveContainer" containerID="08bd17c836a424fd644b4fbfe560085cf2d06aa7c9151a7a2bf0e1d29a30d34a" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.739254 4998 scope.go:117] "RemoveContainer" containerID="1d1830b0a1fe5b88ad88f60a31050e58f9f5b11241b7708d835b4462eedae371" Dec 03 16:08:16 crc kubenswrapper[4998]: I1203 16:08:16.752664 4998 scope.go:117] "RemoveContainer" containerID="af8a03c1a0e3f68bb595014c7e56371b1f9573d230c086e614c3f3d94404140d" Dec 03 16:08:17 crc kubenswrapper[4998]: I1203 16:08:17.684972 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ef027be-08d9-4731-bd35-34f7bb2fe43a" path="/var/lib/kubelet/pods/5ef027be-08d9-4731-bd35-34f7bb2fe43a/volumes" Dec 03 16:08:17 crc kubenswrapper[4998]: I1203 16:08:17.686710 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85eaaffb-675d-4283-9ef9-4765097f8a51" path="/var/lib/kubelet/pods/85eaaffb-675d-4283-9ef9-4765097f8a51/volumes" Dec 03 16:08:17 crc kubenswrapper[4998]: I1203 16:08:17.687780 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea" path="/var/lib/kubelet/pods/8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea/volumes" Dec 03 16:08:17 crc kubenswrapper[4998]: I1203 16:08:17.689293 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c715a96-664c-4b26-8d5a-523a0fb4a17c" path="/var/lib/kubelet/pods/8c715a96-664c-4b26-8d5a-523a0fb4a17c/volumes" Dec 03 16:08:17 crc kubenswrapper[4998]: I1203 16:08:17.689998 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34" path="/var/lib/kubelet/pods/ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34/volumes" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.369847 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-zq6nl" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.431696 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-55d4c"] Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.482629 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rnsnx"] Dec 03 16:08:28 crc kubenswrapper[4998]: E1203 16:08:28.482894 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85eaaffb-675d-4283-9ef9-4765097f8a51" containerName="extract-utilities" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.482912 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="85eaaffb-675d-4283-9ef9-4765097f8a51" containerName="extract-utilities" Dec 03 16:08:28 crc kubenswrapper[4998]: E1203 16:08:28.482923 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ef027be-08d9-4731-bd35-34f7bb2fe43a" containerName="extract-content" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.482931 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ef027be-08d9-4731-bd35-34f7bb2fe43a" containerName="extract-content" Dec 03 16:08:28 crc kubenswrapper[4998]: E1203 16:08:28.482943 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34" containerName="extract-content" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.482953 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34" containerName="extract-content" Dec 03 16:08:28 crc kubenswrapper[4998]: E1203 16:08:28.482963 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea" containerName="registry-server" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.482969 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea" containerName="registry-server" Dec 03 16:08:28 crc kubenswrapper[4998]: E1203 16:08:28.482981 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34" containerName="extract-utilities" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.482989 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34" containerName="extract-utilities" Dec 03 16:08:28 crc kubenswrapper[4998]: E1203 16:08:28.483001 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea" containerName="extract-content" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.483009 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea" containerName="extract-content" Dec 03 16:08:28 crc kubenswrapper[4998]: E1203 16:08:28.483021 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea" containerName="extract-utilities" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.483028 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea" containerName="extract-utilities" Dec 03 16:08:28 crc kubenswrapper[4998]: E1203 16:08:28.483040 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c715a96-664c-4b26-8d5a-523a0fb4a17c" containerName="marketplace-operator" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.483047 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c715a96-664c-4b26-8d5a-523a0fb4a17c" containerName="marketplace-operator" Dec 03 16:08:28 crc kubenswrapper[4998]: E1203 16:08:28.483061 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ef027be-08d9-4731-bd35-34f7bb2fe43a" containerName="extract-utilities" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.483069 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ef027be-08d9-4731-bd35-34f7bb2fe43a" containerName="extract-utilities" Dec 03 16:08:28 crc kubenswrapper[4998]: E1203 16:08:28.483080 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85eaaffb-675d-4283-9ef9-4765097f8a51" containerName="extract-content" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.483089 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="85eaaffb-675d-4283-9ef9-4765097f8a51" containerName="extract-content" Dec 03 16:08:28 crc kubenswrapper[4998]: E1203 16:08:28.483101 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ef027be-08d9-4731-bd35-34f7bb2fe43a" containerName="registry-server" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.483109 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ef027be-08d9-4731-bd35-34f7bb2fe43a" containerName="registry-server" Dec 03 16:08:28 crc kubenswrapper[4998]: E1203 16:08:28.483119 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85eaaffb-675d-4283-9ef9-4765097f8a51" containerName="registry-server" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.483126 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="85eaaffb-675d-4283-9ef9-4765097f8a51" containerName="registry-server" Dec 03 16:08:28 crc kubenswrapper[4998]: E1203 16:08:28.483137 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34" containerName="registry-server" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.483145 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34" containerName="registry-server" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.483248 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bc8f0c4-f1f9-4d6a-9a71-5fa2e36d3cea" containerName="registry-server" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.483264 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea1d0431-ca9a-4275-8ef0-96e9ca7c0e34" containerName="registry-server" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.483279 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="85eaaffb-675d-4283-9ef9-4765097f8a51" containerName="registry-server" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.483303 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ef027be-08d9-4731-bd35-34f7bb2fe43a" containerName="registry-server" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.483312 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c715a96-664c-4b26-8d5a-523a0fb4a17c" containerName="marketplace-operator" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.484178 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rnsnx" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.488743 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.503254 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rnsnx"] Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.650152 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tql8\" (UniqueName: \"kubernetes.io/projected/20cc0757-ee9b-431b-85ea-189a0b33827b-kube-api-access-5tql8\") pod \"redhat-marketplace-rnsnx\" (UID: \"20cc0757-ee9b-431b-85ea-189a0b33827b\") " pod="openshift-marketplace/redhat-marketplace-rnsnx" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.650215 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20cc0757-ee9b-431b-85ea-189a0b33827b-catalog-content\") pod \"redhat-marketplace-rnsnx\" (UID: \"20cc0757-ee9b-431b-85ea-189a0b33827b\") " pod="openshift-marketplace/redhat-marketplace-rnsnx" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.650309 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20cc0757-ee9b-431b-85ea-189a0b33827b-utilities\") pod \"redhat-marketplace-rnsnx\" (UID: \"20cc0757-ee9b-431b-85ea-189a0b33827b\") " pod="openshift-marketplace/redhat-marketplace-rnsnx" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.687239 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jn68h"] Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.696974 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jn68h" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.701808 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.728897 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jn68h"] Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.751449 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20cc0757-ee9b-431b-85ea-189a0b33827b-utilities\") pod \"redhat-marketplace-rnsnx\" (UID: \"20cc0757-ee9b-431b-85ea-189a0b33827b\") " pod="openshift-marketplace/redhat-marketplace-rnsnx" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.751538 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tql8\" (UniqueName: \"kubernetes.io/projected/20cc0757-ee9b-431b-85ea-189a0b33827b-kube-api-access-5tql8\") pod \"redhat-marketplace-rnsnx\" (UID: \"20cc0757-ee9b-431b-85ea-189a0b33827b\") " pod="openshift-marketplace/redhat-marketplace-rnsnx" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.751562 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20cc0757-ee9b-431b-85ea-189a0b33827b-catalog-content\") pod \"redhat-marketplace-rnsnx\" (UID: \"20cc0757-ee9b-431b-85ea-189a0b33827b\") " pod="openshift-marketplace/redhat-marketplace-rnsnx" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.752054 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20cc0757-ee9b-431b-85ea-189a0b33827b-catalog-content\") pod \"redhat-marketplace-rnsnx\" (UID: \"20cc0757-ee9b-431b-85ea-189a0b33827b\") " pod="openshift-marketplace/redhat-marketplace-rnsnx" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.752330 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20cc0757-ee9b-431b-85ea-189a0b33827b-utilities\") pod \"redhat-marketplace-rnsnx\" (UID: \"20cc0757-ee9b-431b-85ea-189a0b33827b\") " pod="openshift-marketplace/redhat-marketplace-rnsnx" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.775621 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tql8\" (UniqueName: \"kubernetes.io/projected/20cc0757-ee9b-431b-85ea-189a0b33827b-kube-api-access-5tql8\") pod \"redhat-marketplace-rnsnx\" (UID: \"20cc0757-ee9b-431b-85ea-189a0b33827b\") " pod="openshift-marketplace/redhat-marketplace-rnsnx" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.802235 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rnsnx" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.852952 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d1c51e3-2a5d-45dc-917f-0101d813f6d6-utilities\") pod \"redhat-operators-jn68h\" (UID: \"5d1c51e3-2a5d-45dc-917f-0101d813f6d6\") " pod="openshift-marketplace/redhat-operators-jn68h" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.853007 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d1c51e3-2a5d-45dc-917f-0101d813f6d6-catalog-content\") pod \"redhat-operators-jn68h\" (UID: \"5d1c51e3-2a5d-45dc-917f-0101d813f6d6\") " pod="openshift-marketplace/redhat-operators-jn68h" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.853067 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvnqw\" (UniqueName: \"kubernetes.io/projected/5d1c51e3-2a5d-45dc-917f-0101d813f6d6-kube-api-access-wvnqw\") pod \"redhat-operators-jn68h\" (UID: \"5d1c51e3-2a5d-45dc-917f-0101d813f6d6\") " pod="openshift-marketplace/redhat-operators-jn68h" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.954352 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvnqw\" (UniqueName: \"kubernetes.io/projected/5d1c51e3-2a5d-45dc-917f-0101d813f6d6-kube-api-access-wvnqw\") pod \"redhat-operators-jn68h\" (UID: \"5d1c51e3-2a5d-45dc-917f-0101d813f6d6\") " pod="openshift-marketplace/redhat-operators-jn68h" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.954462 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d1c51e3-2a5d-45dc-917f-0101d813f6d6-utilities\") pod \"redhat-operators-jn68h\" (UID: \"5d1c51e3-2a5d-45dc-917f-0101d813f6d6\") " pod="openshift-marketplace/redhat-operators-jn68h" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.954498 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d1c51e3-2a5d-45dc-917f-0101d813f6d6-catalog-content\") pod \"redhat-operators-jn68h\" (UID: \"5d1c51e3-2a5d-45dc-917f-0101d813f6d6\") " pod="openshift-marketplace/redhat-operators-jn68h" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.955009 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d1c51e3-2a5d-45dc-917f-0101d813f6d6-catalog-content\") pod \"redhat-operators-jn68h\" (UID: \"5d1c51e3-2a5d-45dc-917f-0101d813f6d6\") " pod="openshift-marketplace/redhat-operators-jn68h" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.955212 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d1c51e3-2a5d-45dc-917f-0101d813f6d6-utilities\") pod \"redhat-operators-jn68h\" (UID: \"5d1c51e3-2a5d-45dc-917f-0101d813f6d6\") " pod="openshift-marketplace/redhat-operators-jn68h" Dec 03 16:08:28 crc kubenswrapper[4998]: I1203 16:08:28.985577 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvnqw\" (UniqueName: \"kubernetes.io/projected/5d1c51e3-2a5d-45dc-917f-0101d813f6d6-kube-api-access-wvnqw\") pod \"redhat-operators-jn68h\" (UID: \"5d1c51e3-2a5d-45dc-917f-0101d813f6d6\") " pod="openshift-marketplace/redhat-operators-jn68h" Dec 03 16:08:29 crc kubenswrapper[4998]: I1203 16:08:29.044816 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jn68h" Dec 03 16:08:29 crc kubenswrapper[4998]: I1203 16:08:29.206212 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rnsnx"] Dec 03 16:08:29 crc kubenswrapper[4998]: W1203 16:08:29.208518 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20cc0757_ee9b_431b_85ea_189a0b33827b.slice/crio-88d6422212911e1c70f49fb75ae27ec881e0746849f4504ddf249e05fa040b08 WatchSource:0}: Error finding container 88d6422212911e1c70f49fb75ae27ec881e0746849f4504ddf249e05fa040b08: Status 404 returned error can't find the container with id 88d6422212911e1c70f49fb75ae27ec881e0746849f4504ddf249e05fa040b08 Dec 03 16:08:29 crc kubenswrapper[4998]: I1203 16:08:29.461730 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jn68h"] Dec 03 16:08:29 crc kubenswrapper[4998]: W1203 16:08:29.468410 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d1c51e3_2a5d_45dc_917f_0101d813f6d6.slice/crio-5b93b2b607f3e5f66022228d5d446a02d927b0b41409b9ddfdc57b04a29cb77c WatchSource:0}: Error finding container 5b93b2b607f3e5f66022228d5d446a02d927b0b41409b9ddfdc57b04a29cb77c: Status 404 returned error can't find the container with id 5b93b2b607f3e5f66022228d5d446a02d927b0b41409b9ddfdc57b04a29cb77c Dec 03 16:08:29 crc kubenswrapper[4998]: I1203 16:08:29.591488 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jn68h" event={"ID":"5d1c51e3-2a5d-45dc-917f-0101d813f6d6","Type":"ContainerStarted","Data":"5b93b2b607f3e5f66022228d5d446a02d927b0b41409b9ddfdc57b04a29cb77c"} Dec 03 16:08:29 crc kubenswrapper[4998]: I1203 16:08:29.593449 4998 generic.go:334] "Generic (PLEG): container finished" podID="20cc0757-ee9b-431b-85ea-189a0b33827b" containerID="f68705cb6b4be49682c6b9f209b9a566f656b5554adba546c124c357854207a2" exitCode=0 Dec 03 16:08:29 crc kubenswrapper[4998]: I1203 16:08:29.593475 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rnsnx" event={"ID":"20cc0757-ee9b-431b-85ea-189a0b33827b","Type":"ContainerDied","Data":"f68705cb6b4be49682c6b9f209b9a566f656b5554adba546c124c357854207a2"} Dec 03 16:08:29 crc kubenswrapper[4998]: I1203 16:08:29.593490 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rnsnx" event={"ID":"20cc0757-ee9b-431b-85ea-189a0b33827b","Type":"ContainerStarted","Data":"88d6422212911e1c70f49fb75ae27ec881e0746849f4504ddf249e05fa040b08"} Dec 03 16:08:30 crc kubenswrapper[4998]: I1203 16:08:30.600851 4998 generic.go:334] "Generic (PLEG): container finished" podID="5d1c51e3-2a5d-45dc-917f-0101d813f6d6" containerID="7de481457ee24970d430730bef1414621d1efccc5bbdd563f50dac0c2c9ac59d" exitCode=0 Dec 03 16:08:30 crc kubenswrapper[4998]: I1203 16:08:30.600951 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jn68h" event={"ID":"5d1c51e3-2a5d-45dc-917f-0101d813f6d6","Type":"ContainerDied","Data":"7de481457ee24970d430730bef1414621d1efccc5bbdd563f50dac0c2c9ac59d"} Dec 03 16:08:30 crc kubenswrapper[4998]: I1203 16:08:30.888692 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tdx5r"] Dec 03 16:08:30 crc kubenswrapper[4998]: I1203 16:08:30.890426 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tdx5r" Dec 03 16:08:30 crc kubenswrapper[4998]: I1203 16:08:30.893068 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 16:08:30 crc kubenswrapper[4998]: I1203 16:08:30.908202 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tdx5r"] Dec 03 16:08:30 crc kubenswrapper[4998]: I1203 16:08:30.979301 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0e26a80-b16f-4dd8-a8b1-c567913e3922-catalog-content\") pod \"certified-operators-tdx5r\" (UID: \"f0e26a80-b16f-4dd8-a8b1-c567913e3922\") " pod="openshift-marketplace/certified-operators-tdx5r" Dec 03 16:08:30 crc kubenswrapper[4998]: I1203 16:08:30.979372 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftptp\" (UniqueName: \"kubernetes.io/projected/f0e26a80-b16f-4dd8-a8b1-c567913e3922-kube-api-access-ftptp\") pod \"certified-operators-tdx5r\" (UID: \"f0e26a80-b16f-4dd8-a8b1-c567913e3922\") " pod="openshift-marketplace/certified-operators-tdx5r" Dec 03 16:08:30 crc kubenswrapper[4998]: I1203 16:08:30.979404 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0e26a80-b16f-4dd8-a8b1-c567913e3922-utilities\") pod \"certified-operators-tdx5r\" (UID: \"f0e26a80-b16f-4dd8-a8b1-c567913e3922\") " pod="openshift-marketplace/certified-operators-tdx5r" Dec 03 16:08:31 crc kubenswrapper[4998]: I1203 16:08:31.080272 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftptp\" (UniqueName: \"kubernetes.io/projected/f0e26a80-b16f-4dd8-a8b1-c567913e3922-kube-api-access-ftptp\") pod \"certified-operators-tdx5r\" (UID: \"f0e26a80-b16f-4dd8-a8b1-c567913e3922\") " pod="openshift-marketplace/certified-operators-tdx5r" Dec 03 16:08:31 crc kubenswrapper[4998]: I1203 16:08:31.080541 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0e26a80-b16f-4dd8-a8b1-c567913e3922-utilities\") pod \"certified-operators-tdx5r\" (UID: \"f0e26a80-b16f-4dd8-a8b1-c567913e3922\") " pod="openshift-marketplace/certified-operators-tdx5r" Dec 03 16:08:31 crc kubenswrapper[4998]: I1203 16:08:31.080596 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0e26a80-b16f-4dd8-a8b1-c567913e3922-catalog-content\") pod \"certified-operators-tdx5r\" (UID: \"f0e26a80-b16f-4dd8-a8b1-c567913e3922\") " pod="openshift-marketplace/certified-operators-tdx5r" Dec 03 16:08:31 crc kubenswrapper[4998]: I1203 16:08:31.081139 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0e26a80-b16f-4dd8-a8b1-c567913e3922-catalog-content\") pod \"certified-operators-tdx5r\" (UID: \"f0e26a80-b16f-4dd8-a8b1-c567913e3922\") " pod="openshift-marketplace/certified-operators-tdx5r" Dec 03 16:08:31 crc kubenswrapper[4998]: I1203 16:08:31.081149 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0e26a80-b16f-4dd8-a8b1-c567913e3922-utilities\") pod \"certified-operators-tdx5r\" (UID: \"f0e26a80-b16f-4dd8-a8b1-c567913e3922\") " pod="openshift-marketplace/certified-operators-tdx5r" Dec 03 16:08:31 crc kubenswrapper[4998]: I1203 16:08:31.094414 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tfv2f"] Dec 03 16:08:31 crc kubenswrapper[4998]: I1203 16:08:31.096526 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tfv2f" Dec 03 16:08:31 crc kubenswrapper[4998]: I1203 16:08:31.098824 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 16:08:31 crc kubenswrapper[4998]: I1203 16:08:31.103024 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tfv2f"] Dec 03 16:08:31 crc kubenswrapper[4998]: I1203 16:08:31.114128 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftptp\" (UniqueName: \"kubernetes.io/projected/f0e26a80-b16f-4dd8-a8b1-c567913e3922-kube-api-access-ftptp\") pod \"certified-operators-tdx5r\" (UID: \"f0e26a80-b16f-4dd8-a8b1-c567913e3922\") " pod="openshift-marketplace/certified-operators-tdx5r" Dec 03 16:08:31 crc kubenswrapper[4998]: I1203 16:08:31.181643 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/816953d5-a5ca-4822-94cb-600d1bbbdb6b-catalog-content\") pod \"community-operators-tfv2f\" (UID: \"816953d5-a5ca-4822-94cb-600d1bbbdb6b\") " pod="openshift-marketplace/community-operators-tfv2f" Dec 03 16:08:31 crc kubenswrapper[4998]: I1203 16:08:31.181986 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgshq\" (UniqueName: \"kubernetes.io/projected/816953d5-a5ca-4822-94cb-600d1bbbdb6b-kube-api-access-wgshq\") pod \"community-operators-tfv2f\" (UID: \"816953d5-a5ca-4822-94cb-600d1bbbdb6b\") " pod="openshift-marketplace/community-operators-tfv2f" Dec 03 16:08:31 crc kubenswrapper[4998]: I1203 16:08:31.182072 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/816953d5-a5ca-4822-94cb-600d1bbbdb6b-utilities\") pod \"community-operators-tfv2f\" (UID: \"816953d5-a5ca-4822-94cb-600d1bbbdb6b\") " pod="openshift-marketplace/community-operators-tfv2f" Dec 03 16:08:31 crc kubenswrapper[4998]: I1203 16:08:31.215612 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tdx5r" Dec 03 16:08:31 crc kubenswrapper[4998]: I1203 16:08:31.284337 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgshq\" (UniqueName: \"kubernetes.io/projected/816953d5-a5ca-4822-94cb-600d1bbbdb6b-kube-api-access-wgshq\") pod \"community-operators-tfv2f\" (UID: \"816953d5-a5ca-4822-94cb-600d1bbbdb6b\") " pod="openshift-marketplace/community-operators-tfv2f" Dec 03 16:08:31 crc kubenswrapper[4998]: I1203 16:08:31.284407 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/816953d5-a5ca-4822-94cb-600d1bbbdb6b-utilities\") pod \"community-operators-tfv2f\" (UID: \"816953d5-a5ca-4822-94cb-600d1bbbdb6b\") " pod="openshift-marketplace/community-operators-tfv2f" Dec 03 16:08:31 crc kubenswrapper[4998]: I1203 16:08:31.284869 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/816953d5-a5ca-4822-94cb-600d1bbbdb6b-utilities\") pod \"community-operators-tfv2f\" (UID: \"816953d5-a5ca-4822-94cb-600d1bbbdb6b\") " pod="openshift-marketplace/community-operators-tfv2f" Dec 03 16:08:31 crc kubenswrapper[4998]: I1203 16:08:31.285146 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/816953d5-a5ca-4822-94cb-600d1bbbdb6b-catalog-content\") pod \"community-operators-tfv2f\" (UID: \"816953d5-a5ca-4822-94cb-600d1bbbdb6b\") " pod="openshift-marketplace/community-operators-tfv2f" Dec 03 16:08:31 crc kubenswrapper[4998]: I1203 16:08:31.285478 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/816953d5-a5ca-4822-94cb-600d1bbbdb6b-catalog-content\") pod \"community-operators-tfv2f\" (UID: \"816953d5-a5ca-4822-94cb-600d1bbbdb6b\") " pod="openshift-marketplace/community-operators-tfv2f" Dec 03 16:08:31 crc kubenswrapper[4998]: I1203 16:08:31.300869 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgshq\" (UniqueName: \"kubernetes.io/projected/816953d5-a5ca-4822-94cb-600d1bbbdb6b-kube-api-access-wgshq\") pod \"community-operators-tfv2f\" (UID: \"816953d5-a5ca-4822-94cb-600d1bbbdb6b\") " pod="openshift-marketplace/community-operators-tfv2f" Dec 03 16:08:31 crc kubenswrapper[4998]: I1203 16:08:31.415256 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tfv2f" Dec 03 16:08:31 crc kubenswrapper[4998]: I1203 16:08:31.627203 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tdx5r"] Dec 03 16:08:31 crc kubenswrapper[4998]: W1203 16:08:31.634096 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf0e26a80_b16f_4dd8_a8b1_c567913e3922.slice/crio-725ba5818c222e6ca31138b6b180b04556224881e5d555dfb64d0c02f8e97a17 WatchSource:0}: Error finding container 725ba5818c222e6ca31138b6b180b04556224881e5d555dfb64d0c02f8e97a17: Status 404 returned error can't find the container with id 725ba5818c222e6ca31138b6b180b04556224881e5d555dfb64d0c02f8e97a17 Dec 03 16:08:31 crc kubenswrapper[4998]: I1203 16:08:31.796918 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tfv2f"] Dec 03 16:08:31 crc kubenswrapper[4998]: W1203 16:08:31.800337 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod816953d5_a5ca_4822_94cb_600d1bbbdb6b.slice/crio-db2c1c6ea6bd617556048d515b79fa2af3379f7119fa1073ff736bf689f435de WatchSource:0}: Error finding container db2c1c6ea6bd617556048d515b79fa2af3379f7119fa1073ff736bf689f435de: Status 404 returned error can't find the container with id db2c1c6ea6bd617556048d515b79fa2af3379f7119fa1073ff736bf689f435de Dec 03 16:08:32 crc kubenswrapper[4998]: I1203 16:08:32.613814 4998 generic.go:334] "Generic (PLEG): container finished" podID="f0e26a80-b16f-4dd8-a8b1-c567913e3922" containerID="6f23123c15f2d4361f98db6590ff9f655d329454b1049a350a72485dd42230be" exitCode=0 Dec 03 16:08:32 crc kubenswrapper[4998]: I1203 16:08:32.614074 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tdx5r" event={"ID":"f0e26a80-b16f-4dd8-a8b1-c567913e3922","Type":"ContainerDied","Data":"6f23123c15f2d4361f98db6590ff9f655d329454b1049a350a72485dd42230be"} Dec 03 16:08:32 crc kubenswrapper[4998]: I1203 16:08:32.614116 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tdx5r" event={"ID":"f0e26a80-b16f-4dd8-a8b1-c567913e3922","Type":"ContainerStarted","Data":"725ba5818c222e6ca31138b6b180b04556224881e5d555dfb64d0c02f8e97a17"} Dec 03 16:08:32 crc kubenswrapper[4998]: I1203 16:08:32.616002 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tfv2f" event={"ID":"816953d5-a5ca-4822-94cb-600d1bbbdb6b","Type":"ContainerStarted","Data":"db2c1c6ea6bd617556048d515b79fa2af3379f7119fa1073ff736bf689f435de"} Dec 03 16:08:34 crc kubenswrapper[4998]: I1203 16:08:34.628997 4998 generic.go:334] "Generic (PLEG): container finished" podID="20cc0757-ee9b-431b-85ea-189a0b33827b" containerID="fa5fdec13bc2ca7b83daed27c1c8a81b384149c36746d775587557b04d9420e6" exitCode=0 Dec 03 16:08:34 crc kubenswrapper[4998]: I1203 16:08:34.629082 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rnsnx" event={"ID":"20cc0757-ee9b-431b-85ea-189a0b33827b","Type":"ContainerDied","Data":"fa5fdec13bc2ca7b83daed27c1c8a81b384149c36746d775587557b04d9420e6"} Dec 03 16:08:34 crc kubenswrapper[4998]: I1203 16:08:34.632773 4998 generic.go:334] "Generic (PLEG): container finished" podID="f0e26a80-b16f-4dd8-a8b1-c567913e3922" containerID="f26e8663e03fd7cc3c45566546c95f8832c8fcc1a18c0a08faa6a8e76568ec21" exitCode=0 Dec 03 16:08:34 crc kubenswrapper[4998]: I1203 16:08:34.632840 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tdx5r" event={"ID":"f0e26a80-b16f-4dd8-a8b1-c567913e3922","Type":"ContainerDied","Data":"f26e8663e03fd7cc3c45566546c95f8832c8fcc1a18c0a08faa6a8e76568ec21"} Dec 03 16:08:34 crc kubenswrapper[4998]: I1203 16:08:34.636578 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jn68h" event={"ID":"5d1c51e3-2a5d-45dc-917f-0101d813f6d6","Type":"ContainerStarted","Data":"79f8035881054747209e109657f689908bbb5e91cbecf01de567c09f178de58d"} Dec 03 16:08:34 crc kubenswrapper[4998]: I1203 16:08:34.639210 4998 generic.go:334] "Generic (PLEG): container finished" podID="816953d5-a5ca-4822-94cb-600d1bbbdb6b" containerID="adadcb4a6a35fa670d260da850a459953da744421077c24879c8a32cf37f0409" exitCode=0 Dec 03 16:08:34 crc kubenswrapper[4998]: I1203 16:08:34.639270 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tfv2f" event={"ID":"816953d5-a5ca-4822-94cb-600d1bbbdb6b","Type":"ContainerDied","Data":"adadcb4a6a35fa670d260da850a459953da744421077c24879c8a32cf37f0409"} Dec 03 16:08:35 crc kubenswrapper[4998]: I1203 16:08:35.647107 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tdx5r" event={"ID":"f0e26a80-b16f-4dd8-a8b1-c567913e3922","Type":"ContainerStarted","Data":"6372b015c6684e998aa5a52ef875e9f2852fff9c50894f650f9dbdf6558e0e2f"} Dec 03 16:08:35 crc kubenswrapper[4998]: I1203 16:08:35.650420 4998 generic.go:334] "Generic (PLEG): container finished" podID="5d1c51e3-2a5d-45dc-917f-0101d813f6d6" containerID="79f8035881054747209e109657f689908bbb5e91cbecf01de567c09f178de58d" exitCode=0 Dec 03 16:08:35 crc kubenswrapper[4998]: I1203 16:08:35.650505 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jn68h" event={"ID":"5d1c51e3-2a5d-45dc-917f-0101d813f6d6","Type":"ContainerDied","Data":"79f8035881054747209e109657f689908bbb5e91cbecf01de567c09f178de58d"} Dec 03 16:08:35 crc kubenswrapper[4998]: I1203 16:08:35.657207 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tfv2f" event={"ID":"816953d5-a5ca-4822-94cb-600d1bbbdb6b","Type":"ContainerStarted","Data":"251c17f3e990734bc0e154eaaf214c1f5d15b8db902b3a9dc2903061c48c64c3"} Dec 03 16:08:35 crc kubenswrapper[4998]: I1203 16:08:35.660262 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rnsnx" event={"ID":"20cc0757-ee9b-431b-85ea-189a0b33827b","Type":"ContainerStarted","Data":"968c8c86140fd812b149542a6860c9cf08378486b903a4534b30fcfacf8f2deb"} Dec 03 16:08:35 crc kubenswrapper[4998]: I1203 16:08:35.676137 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tdx5r" podStartSLOduration=3.1723797559999998 podStartE2EDuration="5.676114032s" podCreationTimestamp="2025-12-03 16:08:30 +0000 UTC" firstStartedPulling="2025-12-03 16:08:32.644443164 +0000 UTC m=+291.256143427" lastFinishedPulling="2025-12-03 16:08:35.14817746 +0000 UTC m=+293.759877703" observedRunningTime="2025-12-03 16:08:35.672029253 +0000 UTC m=+294.283729486" watchObservedRunningTime="2025-12-03 16:08:35.676114032 +0000 UTC m=+294.287814265" Dec 03 16:08:35 crc kubenswrapper[4998]: I1203 16:08:35.697999 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rnsnx" podStartSLOduration=3.191137707 podStartE2EDuration="7.697974288s" podCreationTimestamp="2025-12-03 16:08:28 +0000 UTC" firstStartedPulling="2025-12-03 16:08:30.60258077 +0000 UTC m=+289.214281033" lastFinishedPulling="2025-12-03 16:08:35.109417381 +0000 UTC m=+293.721117614" observedRunningTime="2025-12-03 16:08:35.69616561 +0000 UTC m=+294.307865833" watchObservedRunningTime="2025-12-03 16:08:35.697974288 +0000 UTC m=+294.309674531" Dec 03 16:08:36 crc kubenswrapper[4998]: I1203 16:08:36.669773 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jn68h" event={"ID":"5d1c51e3-2a5d-45dc-917f-0101d813f6d6","Type":"ContainerStarted","Data":"a69dca292a00a6e42f5a3a9869bcb7e73d4c33d6fe98d8fa0d8a8d6559feb829"} Dec 03 16:08:36 crc kubenswrapper[4998]: I1203 16:08:36.672230 4998 generic.go:334] "Generic (PLEG): container finished" podID="816953d5-a5ca-4822-94cb-600d1bbbdb6b" containerID="251c17f3e990734bc0e154eaaf214c1f5d15b8db902b3a9dc2903061c48c64c3" exitCode=0 Dec 03 16:08:36 crc kubenswrapper[4998]: I1203 16:08:36.672283 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tfv2f" event={"ID":"816953d5-a5ca-4822-94cb-600d1bbbdb6b","Type":"ContainerDied","Data":"251c17f3e990734bc0e154eaaf214c1f5d15b8db902b3a9dc2903061c48c64c3"} Dec 03 16:08:36 crc kubenswrapper[4998]: I1203 16:08:36.695973 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jn68h" podStartSLOduration=3.226079554 podStartE2EDuration="8.695955141s" podCreationTimestamp="2025-12-03 16:08:28 +0000 UTC" firstStartedPulling="2025-12-03 16:08:30.602278722 +0000 UTC m=+289.213978965" lastFinishedPulling="2025-12-03 16:08:36.072154329 +0000 UTC m=+294.683854552" observedRunningTime="2025-12-03 16:08:36.692420576 +0000 UTC m=+295.304120809" watchObservedRunningTime="2025-12-03 16:08:36.695955141 +0000 UTC m=+295.307655364" Dec 03 16:08:37 crc kubenswrapper[4998]: I1203 16:08:37.684875 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tfv2f" event={"ID":"816953d5-a5ca-4822-94cb-600d1bbbdb6b","Type":"ContainerStarted","Data":"e63a0b66f917696b8b700cfb4f71a25bc43b3c43d5a39a4fbe5e36a7a33d867b"} Dec 03 16:08:37 crc kubenswrapper[4998]: I1203 16:08:37.701005 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tfv2f" podStartSLOduration=4.286072908 podStartE2EDuration="6.700980362s" podCreationTimestamp="2025-12-03 16:08:31 +0000 UTC" firstStartedPulling="2025-12-03 16:08:34.643367979 +0000 UTC m=+293.255068212" lastFinishedPulling="2025-12-03 16:08:37.058275443 +0000 UTC m=+295.669975666" observedRunningTime="2025-12-03 16:08:37.700482099 +0000 UTC m=+296.312182362" watchObservedRunningTime="2025-12-03 16:08:37.700980362 +0000 UTC m=+296.312680595" Dec 03 16:08:38 crc kubenswrapper[4998]: I1203 16:08:38.803387 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rnsnx" Dec 03 16:08:38 crc kubenswrapper[4998]: I1203 16:08:38.803469 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rnsnx" Dec 03 16:08:38 crc kubenswrapper[4998]: I1203 16:08:38.852058 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rnsnx" Dec 03 16:08:39 crc kubenswrapper[4998]: I1203 16:08:39.045431 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jn68h" Dec 03 16:08:39 crc kubenswrapper[4998]: I1203 16:08:39.047802 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jn68h" Dec 03 16:08:40 crc kubenswrapper[4998]: I1203 16:08:40.094098 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jn68h" podUID="5d1c51e3-2a5d-45dc-917f-0101d813f6d6" containerName="registry-server" probeResult="failure" output=< Dec 03 16:08:40 crc kubenswrapper[4998]: timeout: failed to connect service ":50051" within 1s Dec 03 16:08:40 crc kubenswrapper[4998]: > Dec 03 16:08:41 crc kubenswrapper[4998]: I1203 16:08:41.216129 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tdx5r" Dec 03 16:08:41 crc kubenswrapper[4998]: I1203 16:08:41.216519 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tdx5r" Dec 03 16:08:41 crc kubenswrapper[4998]: I1203 16:08:41.255899 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tdx5r" Dec 03 16:08:41 crc kubenswrapper[4998]: I1203 16:08:41.416796 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tfv2f" Dec 03 16:08:41 crc kubenswrapper[4998]: I1203 16:08:41.416875 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tfv2f" Dec 03 16:08:41 crc kubenswrapper[4998]: I1203 16:08:41.456676 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tfv2f" Dec 03 16:08:41 crc kubenswrapper[4998]: I1203 16:08:41.760896 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tdx5r" Dec 03 16:08:45 crc kubenswrapper[4998]: I1203 16:08:45.013727 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz"] Dec 03 16:08:45 crc kubenswrapper[4998]: I1203 16:08:45.014254 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz" podUID="1ae0bcea-dffa-471b-84eb-55298fbd29b5" containerName="route-controller-manager" containerID="cri-o://f8dbf1581fc65eacd6eb521d63a265d6647da761e10618c9362bb2f2bac0113d" gracePeriod=30 Dec 03 16:08:46 crc kubenswrapper[4998]: I1203 16:08:46.759884 4998 patch_prober.go:28] interesting pod/route-controller-manager-59877c47b5-j99tz container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" start-of-body= Dec 03 16:08:46 crc kubenswrapper[4998]: I1203 16:08:46.759972 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz" podUID="1ae0bcea-dffa-471b-84eb-55298fbd29b5" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" Dec 03 16:08:48 crc kubenswrapper[4998]: I1203 16:08:48.877712 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rnsnx" Dec 03 16:08:49 crc kubenswrapper[4998]: I1203 16:08:49.112529 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jn68h" Dec 03 16:08:49 crc kubenswrapper[4998]: I1203 16:08:49.172832 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jn68h" Dec 03 16:08:51 crc kubenswrapper[4998]: I1203 16:08:51.452567 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tfv2f" Dec 03 16:08:52 crc kubenswrapper[4998]: I1203 16:08:52.882260 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz" Dec 03 16:08:52 crc kubenswrapper[4998]: I1203 16:08:52.906257 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7dcb4cd7b8-hc9rp"] Dec 03 16:08:52 crc kubenswrapper[4998]: E1203 16:08:52.906471 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ae0bcea-dffa-471b-84eb-55298fbd29b5" containerName="route-controller-manager" Dec 03 16:08:52 crc kubenswrapper[4998]: I1203 16:08:52.906484 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ae0bcea-dffa-471b-84eb-55298fbd29b5" containerName="route-controller-manager" Dec 03 16:08:52 crc kubenswrapper[4998]: I1203 16:08:52.906594 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ae0bcea-dffa-471b-84eb-55298fbd29b5" containerName="route-controller-manager" Dec 03 16:08:52 crc kubenswrapper[4998]: I1203 16:08:52.907109 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7dcb4cd7b8-hc9rp" Dec 03 16:08:52 crc kubenswrapper[4998]: I1203 16:08:52.929724 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7dcb4cd7b8-hc9rp"] Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.004153 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ae0bcea-dffa-471b-84eb-55298fbd29b5-config\") pod \"1ae0bcea-dffa-471b-84eb-55298fbd29b5\" (UID: \"1ae0bcea-dffa-471b-84eb-55298fbd29b5\") " Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.004209 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhjgc\" (UniqueName: \"kubernetes.io/projected/1ae0bcea-dffa-471b-84eb-55298fbd29b5-kube-api-access-rhjgc\") pod \"1ae0bcea-dffa-471b-84eb-55298fbd29b5\" (UID: \"1ae0bcea-dffa-471b-84eb-55298fbd29b5\") " Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.004272 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1ae0bcea-dffa-471b-84eb-55298fbd29b5-client-ca\") pod \"1ae0bcea-dffa-471b-84eb-55298fbd29b5\" (UID: \"1ae0bcea-dffa-471b-84eb-55298fbd29b5\") " Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.004308 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ae0bcea-dffa-471b-84eb-55298fbd29b5-serving-cert\") pod \"1ae0bcea-dffa-471b-84eb-55298fbd29b5\" (UID: \"1ae0bcea-dffa-471b-84eb-55298fbd29b5\") " Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.004505 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hts9\" (UniqueName: \"kubernetes.io/projected/0457528d-8655-44de-a2d4-ff47a39d40dd-kube-api-access-5hts9\") pod \"route-controller-manager-7dcb4cd7b8-hc9rp\" (UID: \"0457528d-8655-44de-a2d4-ff47a39d40dd\") " pod="openshift-route-controller-manager/route-controller-manager-7dcb4cd7b8-hc9rp" Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.004539 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0457528d-8655-44de-a2d4-ff47a39d40dd-serving-cert\") pod \"route-controller-manager-7dcb4cd7b8-hc9rp\" (UID: \"0457528d-8655-44de-a2d4-ff47a39d40dd\") " pod="openshift-route-controller-manager/route-controller-manager-7dcb4cd7b8-hc9rp" Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.004568 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0457528d-8655-44de-a2d4-ff47a39d40dd-client-ca\") pod \"route-controller-manager-7dcb4cd7b8-hc9rp\" (UID: \"0457528d-8655-44de-a2d4-ff47a39d40dd\") " pod="openshift-route-controller-manager/route-controller-manager-7dcb4cd7b8-hc9rp" Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.004694 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0457528d-8655-44de-a2d4-ff47a39d40dd-config\") pod \"route-controller-manager-7dcb4cd7b8-hc9rp\" (UID: \"0457528d-8655-44de-a2d4-ff47a39d40dd\") " pod="openshift-route-controller-manager/route-controller-manager-7dcb4cd7b8-hc9rp" Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.005152 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ae0bcea-dffa-471b-84eb-55298fbd29b5-config" (OuterVolumeSpecName: "config") pod "1ae0bcea-dffa-471b-84eb-55298fbd29b5" (UID: "1ae0bcea-dffa-471b-84eb-55298fbd29b5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.005184 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ae0bcea-dffa-471b-84eb-55298fbd29b5-client-ca" (OuterVolumeSpecName: "client-ca") pod "1ae0bcea-dffa-471b-84eb-55298fbd29b5" (UID: "1ae0bcea-dffa-471b-84eb-55298fbd29b5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.010463 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ae0bcea-dffa-471b-84eb-55298fbd29b5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1ae0bcea-dffa-471b-84eb-55298fbd29b5" (UID: "1ae0bcea-dffa-471b-84eb-55298fbd29b5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.010897 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ae0bcea-dffa-471b-84eb-55298fbd29b5-kube-api-access-rhjgc" (OuterVolumeSpecName: "kube-api-access-rhjgc") pod "1ae0bcea-dffa-471b-84eb-55298fbd29b5" (UID: "1ae0bcea-dffa-471b-84eb-55298fbd29b5"). InnerVolumeSpecName "kube-api-access-rhjgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.105807 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0457528d-8655-44de-a2d4-ff47a39d40dd-client-ca\") pod \"route-controller-manager-7dcb4cd7b8-hc9rp\" (UID: \"0457528d-8655-44de-a2d4-ff47a39d40dd\") " pod="openshift-route-controller-manager/route-controller-manager-7dcb4cd7b8-hc9rp" Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.105935 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0457528d-8655-44de-a2d4-ff47a39d40dd-config\") pod \"route-controller-manager-7dcb4cd7b8-hc9rp\" (UID: \"0457528d-8655-44de-a2d4-ff47a39d40dd\") " pod="openshift-route-controller-manager/route-controller-manager-7dcb4cd7b8-hc9rp" Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.106098 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hts9\" (UniqueName: \"kubernetes.io/projected/0457528d-8655-44de-a2d4-ff47a39d40dd-kube-api-access-5hts9\") pod \"route-controller-manager-7dcb4cd7b8-hc9rp\" (UID: \"0457528d-8655-44de-a2d4-ff47a39d40dd\") " pod="openshift-route-controller-manager/route-controller-manager-7dcb4cd7b8-hc9rp" Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.106190 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0457528d-8655-44de-a2d4-ff47a39d40dd-serving-cert\") pod \"route-controller-manager-7dcb4cd7b8-hc9rp\" (UID: \"0457528d-8655-44de-a2d4-ff47a39d40dd\") " pod="openshift-route-controller-manager/route-controller-manager-7dcb4cd7b8-hc9rp" Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.106257 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ae0bcea-dffa-471b-84eb-55298fbd29b5-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.106281 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhjgc\" (UniqueName: \"kubernetes.io/projected/1ae0bcea-dffa-471b-84eb-55298fbd29b5-kube-api-access-rhjgc\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.106300 4998 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1ae0bcea-dffa-471b-84eb-55298fbd29b5-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.106316 4998 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ae0bcea-dffa-471b-84eb-55298fbd29b5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.107960 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0457528d-8655-44de-a2d4-ff47a39d40dd-client-ca\") pod \"route-controller-manager-7dcb4cd7b8-hc9rp\" (UID: \"0457528d-8655-44de-a2d4-ff47a39d40dd\") " pod="openshift-route-controller-manager/route-controller-manager-7dcb4cd7b8-hc9rp" Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.108680 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0457528d-8655-44de-a2d4-ff47a39d40dd-config\") pod \"route-controller-manager-7dcb4cd7b8-hc9rp\" (UID: \"0457528d-8655-44de-a2d4-ff47a39d40dd\") " pod="openshift-route-controller-manager/route-controller-manager-7dcb4cd7b8-hc9rp" Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.112261 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0457528d-8655-44de-a2d4-ff47a39d40dd-serving-cert\") pod \"route-controller-manager-7dcb4cd7b8-hc9rp\" (UID: \"0457528d-8655-44de-a2d4-ff47a39d40dd\") " pod="openshift-route-controller-manager/route-controller-manager-7dcb4cd7b8-hc9rp" Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.136951 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hts9\" (UniqueName: \"kubernetes.io/projected/0457528d-8655-44de-a2d4-ff47a39d40dd-kube-api-access-5hts9\") pod \"route-controller-manager-7dcb4cd7b8-hc9rp\" (UID: \"0457528d-8655-44de-a2d4-ff47a39d40dd\") " pod="openshift-route-controller-manager/route-controller-manager-7dcb4cd7b8-hc9rp" Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.222125 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7dcb4cd7b8-hc9rp" Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.503482 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" podUID="e6207e94-fd68-464e-9830-895d96a3437c" containerName="registry" containerID="cri-o://70535b37817c6a127f7e248412618455c6549628e70e6a2ff99e6080d32be1a0" gracePeriod=30 Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.509332 4998 generic.go:334] "Generic (PLEG): container finished" podID="1ae0bcea-dffa-471b-84eb-55298fbd29b5" containerID="f8dbf1581fc65eacd6eb521d63a265d6647da761e10618c9362bb2f2bac0113d" exitCode=0 Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.509433 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz" event={"ID":"1ae0bcea-dffa-471b-84eb-55298fbd29b5","Type":"ContainerDied","Data":"f8dbf1581fc65eacd6eb521d63a265d6647da761e10618c9362bb2f2bac0113d"} Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.509505 4998 scope.go:117] "RemoveContainer" containerID="f8dbf1581fc65eacd6eb521d63a265d6647da761e10618c9362bb2f2bac0113d" Dec 03 16:08:53 crc kubenswrapper[4998]: I1203 16:08:53.718423 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7dcb4cd7b8-hc9rp"] Dec 03 16:08:53 crc kubenswrapper[4998]: W1203 16:08:53.723986 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0457528d_8655_44de_a2d4_ff47a39d40dd.slice/crio-c64a561fb082ee15fe1adb0b5226f4a46acef76bb1a09a414136acc131403910 WatchSource:0}: Error finding container c64a561fb082ee15fe1adb0b5226f4a46acef76bb1a09a414136acc131403910: Status 404 returned error can't find the container with id c64a561fb082ee15fe1adb0b5226f4a46acef76bb1a09a414136acc131403910 Dec 03 16:08:54 crc kubenswrapper[4998]: I1203 16:08:54.518532 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7dcb4cd7b8-hc9rp" event={"ID":"0457528d-8655-44de-a2d4-ff47a39d40dd","Type":"ContainerStarted","Data":"c64a561fb082ee15fe1adb0b5226f4a46acef76bb1a09a414136acc131403910"} Dec 03 16:08:54 crc kubenswrapper[4998]: I1203 16:08:54.521255 4998 generic.go:334] "Generic (PLEG): container finished" podID="e6207e94-fd68-464e-9830-895d96a3437c" containerID="70535b37817c6a127f7e248412618455c6549628e70e6a2ff99e6080d32be1a0" exitCode=0 Dec 03 16:08:54 crc kubenswrapper[4998]: I1203 16:08:54.521380 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" event={"ID":"e6207e94-fd68-464e-9830-895d96a3437c","Type":"ContainerDied","Data":"70535b37817c6a127f7e248412618455c6549628e70e6a2ff99e6080d32be1a0"} Dec 03 16:08:54 crc kubenswrapper[4998]: I1203 16:08:54.523406 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz" event={"ID":"1ae0bcea-dffa-471b-84eb-55298fbd29b5","Type":"ContainerDied","Data":"839dd84a48b98fedb523c2a1e7ad25425781a276cfd3b7c750ba8fb2f292e6c3"} Dec 03 16:08:54 crc kubenswrapper[4998]: I1203 16:08:54.523538 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz" Dec 03 16:08:54 crc kubenswrapper[4998]: I1203 16:08:54.549965 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz"] Dec 03 16:08:54 crc kubenswrapper[4998]: I1203 16:08:54.555788 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59877c47b5-j99tz"] Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.170518 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.335627 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e6207e94-fd68-464e-9830-895d96a3437c-registry-tls\") pod \"e6207e94-fd68-464e-9830-895d96a3437c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.335724 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e6207e94-fd68-464e-9830-895d96a3437c-installation-pull-secrets\") pod \"e6207e94-fd68-464e-9830-895d96a3437c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.335823 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e6207e94-fd68-464e-9830-895d96a3437c-registry-certificates\") pod \"e6207e94-fd68-464e-9830-895d96a3437c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.335864 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e6207e94-fd68-464e-9830-895d96a3437c-bound-sa-token\") pod \"e6207e94-fd68-464e-9830-895d96a3437c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.336076 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"e6207e94-fd68-464e-9830-895d96a3437c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.336138 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w56vz\" (UniqueName: \"kubernetes.io/projected/e6207e94-fd68-464e-9830-895d96a3437c-kube-api-access-w56vz\") pod \"e6207e94-fd68-464e-9830-895d96a3437c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.336217 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e6207e94-fd68-464e-9830-895d96a3437c-ca-trust-extracted\") pod \"e6207e94-fd68-464e-9830-895d96a3437c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.336321 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e6207e94-fd68-464e-9830-895d96a3437c-trusted-ca\") pod \"e6207e94-fd68-464e-9830-895d96a3437c\" (UID: \"e6207e94-fd68-464e-9830-895d96a3437c\") " Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.338202 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6207e94-fd68-464e-9830-895d96a3437c-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "e6207e94-fd68-464e-9830-895d96a3437c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.343108 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6207e94-fd68-464e-9830-895d96a3437c-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "e6207e94-fd68-464e-9830-895d96a3437c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.344806 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6207e94-fd68-464e-9830-895d96a3437c-kube-api-access-w56vz" (OuterVolumeSpecName: "kube-api-access-w56vz") pod "e6207e94-fd68-464e-9830-895d96a3437c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c"). InnerVolumeSpecName "kube-api-access-w56vz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.345057 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6207e94-fd68-464e-9830-895d96a3437c-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "e6207e94-fd68-464e-9830-895d96a3437c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.346388 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6207e94-fd68-464e-9830-895d96a3437c-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "e6207e94-fd68-464e-9830-895d96a3437c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.346516 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "e6207e94-fd68-464e-9830-895d96a3437c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.351363 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6207e94-fd68-464e-9830-895d96a3437c-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "e6207e94-fd68-464e-9830-895d96a3437c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.369426 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6207e94-fd68-464e-9830-895d96a3437c-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "e6207e94-fd68-464e-9830-895d96a3437c" (UID: "e6207e94-fd68-464e-9830-895d96a3437c"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.437995 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w56vz\" (UniqueName: \"kubernetes.io/projected/e6207e94-fd68-464e-9830-895d96a3437c-kube-api-access-w56vz\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.438031 4998 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e6207e94-fd68-464e-9830-895d96a3437c-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.438041 4998 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e6207e94-fd68-464e-9830-895d96a3437c-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.438054 4998 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e6207e94-fd68-464e-9830-895d96a3437c-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.438062 4998 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e6207e94-fd68-464e-9830-895d96a3437c-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.438070 4998 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e6207e94-fd68-464e-9830-895d96a3437c-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.438078 4998 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e6207e94-fd68-464e-9830-895d96a3437c-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.529989 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7dcb4cd7b8-hc9rp" event={"ID":"0457528d-8655-44de-a2d4-ff47a39d40dd","Type":"ContainerStarted","Data":"d735652dc2f2f465c3a3415d33ab8f13bf98d4963a59a60f8d5bff529184601c"} Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.531675 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7dcb4cd7b8-hc9rp" Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.533373 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" event={"ID":"e6207e94-fd68-464e-9830-895d96a3437c","Type":"ContainerDied","Data":"5b11427285b9f80ee0e5488edfadb6c4e519e42ce1398edc3f8c155c201988a3"} Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.533426 4998 scope.go:117] "RemoveContainer" containerID="70535b37817c6a127f7e248412618455c6549628e70e6a2ff99e6080d32be1a0" Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.533656 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-55d4c" Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.535804 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7dcb4cd7b8-hc9rp" Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.566907 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7dcb4cd7b8-hc9rp" podStartSLOduration=10.56688583 podStartE2EDuration="10.56688583s" podCreationTimestamp="2025-12-03 16:08:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:08:55.561707812 +0000 UTC m=+314.173408045" watchObservedRunningTime="2025-12-03 16:08:55.56688583 +0000 UTC m=+314.178586063" Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.604237 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-55d4c"] Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.606885 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-55d4c"] Dec 03 16:08:55 crc kubenswrapper[4998]: E1203 16:08:55.611930 4998 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6207e94_fd68_464e_9830_895d96a3437c.slice\": RecentStats: unable to find data in memory cache]" Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.685095 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ae0bcea-dffa-471b-84eb-55298fbd29b5" path="/var/lib/kubelet/pods/1ae0bcea-dffa-471b-84eb-55298fbd29b5/volumes" Dec 03 16:08:55 crc kubenswrapper[4998]: I1203 16:08:55.685702 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6207e94-fd68-464e-9830-895d96a3437c" path="/var/lib/kubelet/pods/e6207e94-fd68-464e-9830-895d96a3437c/volumes" Dec 03 16:09:27 crc kubenswrapper[4998]: I1203 16:09:27.111718 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:09:27 crc kubenswrapper[4998]: I1203 16:09:27.112868 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:09:57 crc kubenswrapper[4998]: I1203 16:09:57.111021 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:09:57 crc kubenswrapper[4998]: I1203 16:09:57.111523 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:10:27 crc kubenswrapper[4998]: I1203 16:10:27.111284 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:10:27 crc kubenswrapper[4998]: I1203 16:10:27.111914 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:10:27 crc kubenswrapper[4998]: I1203 16:10:27.111970 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:10:27 crc kubenswrapper[4998]: I1203 16:10:27.112792 4998 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4005a05bf5e77097685a266b8f7990957d3876e9377b8eac327681f73b327265"} pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 16:10:27 crc kubenswrapper[4998]: I1203 16:10:27.112862 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" containerID="cri-o://4005a05bf5e77097685a266b8f7990957d3876e9377b8eac327681f73b327265" gracePeriod=600 Dec 03 16:10:28 crc kubenswrapper[4998]: I1203 16:10:28.042324 4998 generic.go:334] "Generic (PLEG): container finished" podID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerID="4005a05bf5e77097685a266b8f7990957d3876e9377b8eac327681f73b327265" exitCode=0 Dec 03 16:10:28 crc kubenswrapper[4998]: I1203 16:10:28.042420 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerDied","Data":"4005a05bf5e77097685a266b8f7990957d3876e9377b8eac327681f73b327265"} Dec 03 16:10:28 crc kubenswrapper[4998]: I1203 16:10:28.042590 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"947673c25a57749d81b7b602dfa5d8e08e3e29d4682222a096f3e504d993caf4"} Dec 03 16:10:28 crc kubenswrapper[4998]: I1203 16:10:28.042622 4998 scope.go:117] "RemoveContainer" containerID="7df0f32828517f262cce50b06429b2f45dc22ed31466e1713ef858018738e4a8" Dec 03 16:12:27 crc kubenswrapper[4998]: I1203 16:12:27.111550 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:12:27 crc kubenswrapper[4998]: I1203 16:12:27.112137 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:12:57 crc kubenswrapper[4998]: I1203 16:12:57.111430 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:12:57 crc kubenswrapper[4998]: I1203 16:12:57.113077 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:13:27 crc kubenswrapper[4998]: I1203 16:13:27.111252 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:13:27 crc kubenswrapper[4998]: I1203 16:13:27.111797 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:13:27 crc kubenswrapper[4998]: I1203 16:13:27.111851 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:13:27 crc kubenswrapper[4998]: I1203 16:13:27.112433 4998 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"947673c25a57749d81b7b602dfa5d8e08e3e29d4682222a096f3e504d993caf4"} pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 16:13:27 crc kubenswrapper[4998]: I1203 16:13:27.112484 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" containerID="cri-o://947673c25a57749d81b7b602dfa5d8e08e3e29d4682222a096f3e504d993caf4" gracePeriod=600 Dec 03 16:13:28 crc kubenswrapper[4998]: I1203 16:13:28.186451 4998 generic.go:334] "Generic (PLEG): container finished" podID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerID="947673c25a57749d81b7b602dfa5d8e08e3e29d4682222a096f3e504d993caf4" exitCode=0 Dec 03 16:13:28 crc kubenswrapper[4998]: I1203 16:13:28.186537 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerDied","Data":"947673c25a57749d81b7b602dfa5d8e08e3e29d4682222a096f3e504d993caf4"} Dec 03 16:13:28 crc kubenswrapper[4998]: I1203 16:13:28.186744 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"bc9f09c91a152bc17df6a34b6b589619602b656e78ba0e9011a364c51abd030e"} Dec 03 16:13:28 crc kubenswrapper[4998]: I1203 16:13:28.186789 4998 scope.go:117] "RemoveContainer" containerID="4005a05bf5e77097685a266b8f7990957d3876e9377b8eac327681f73b327265" Dec 03 16:15:00 crc kubenswrapper[4998]: I1203 16:15:00.171725 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412975-hfbt9"] Dec 03 16:15:00 crc kubenswrapper[4998]: E1203 16:15:00.172531 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6207e94-fd68-464e-9830-895d96a3437c" containerName="registry" Dec 03 16:15:00 crc kubenswrapper[4998]: I1203 16:15:00.172544 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6207e94-fd68-464e-9830-895d96a3437c" containerName="registry" Dec 03 16:15:00 crc kubenswrapper[4998]: I1203 16:15:00.172641 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6207e94-fd68-464e-9830-895d96a3437c" containerName="registry" Dec 03 16:15:00 crc kubenswrapper[4998]: I1203 16:15:00.173017 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412975-hfbt9" Dec 03 16:15:00 crc kubenswrapper[4998]: I1203 16:15:00.174896 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 16:15:00 crc kubenswrapper[4998]: I1203 16:15:00.175531 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 16:15:00 crc kubenswrapper[4998]: I1203 16:15:00.185735 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412975-hfbt9"] Dec 03 16:15:00 crc kubenswrapper[4998]: I1203 16:15:00.229301 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtgwr\" (UniqueName: \"kubernetes.io/projected/f249b9c6-1ad9-44b0-8475-8c11dd212f01-kube-api-access-gtgwr\") pod \"collect-profiles-29412975-hfbt9\" (UID: \"f249b9c6-1ad9-44b0-8475-8c11dd212f01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412975-hfbt9" Dec 03 16:15:00 crc kubenswrapper[4998]: I1203 16:15:00.229469 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f249b9c6-1ad9-44b0-8475-8c11dd212f01-config-volume\") pod \"collect-profiles-29412975-hfbt9\" (UID: \"f249b9c6-1ad9-44b0-8475-8c11dd212f01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412975-hfbt9" Dec 03 16:15:00 crc kubenswrapper[4998]: I1203 16:15:00.229699 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f249b9c6-1ad9-44b0-8475-8c11dd212f01-secret-volume\") pod \"collect-profiles-29412975-hfbt9\" (UID: \"f249b9c6-1ad9-44b0-8475-8c11dd212f01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412975-hfbt9" Dec 03 16:15:00 crc kubenswrapper[4998]: I1203 16:15:00.331220 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f249b9c6-1ad9-44b0-8475-8c11dd212f01-secret-volume\") pod \"collect-profiles-29412975-hfbt9\" (UID: \"f249b9c6-1ad9-44b0-8475-8c11dd212f01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412975-hfbt9" Dec 03 16:15:00 crc kubenswrapper[4998]: I1203 16:15:00.331331 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtgwr\" (UniqueName: \"kubernetes.io/projected/f249b9c6-1ad9-44b0-8475-8c11dd212f01-kube-api-access-gtgwr\") pod \"collect-profiles-29412975-hfbt9\" (UID: \"f249b9c6-1ad9-44b0-8475-8c11dd212f01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412975-hfbt9" Dec 03 16:15:00 crc kubenswrapper[4998]: I1203 16:15:00.331401 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f249b9c6-1ad9-44b0-8475-8c11dd212f01-config-volume\") pod \"collect-profiles-29412975-hfbt9\" (UID: \"f249b9c6-1ad9-44b0-8475-8c11dd212f01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412975-hfbt9" Dec 03 16:15:00 crc kubenswrapper[4998]: I1203 16:15:00.333062 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f249b9c6-1ad9-44b0-8475-8c11dd212f01-config-volume\") pod \"collect-profiles-29412975-hfbt9\" (UID: \"f249b9c6-1ad9-44b0-8475-8c11dd212f01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412975-hfbt9" Dec 03 16:15:00 crc kubenswrapper[4998]: I1203 16:15:00.339529 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f249b9c6-1ad9-44b0-8475-8c11dd212f01-secret-volume\") pod \"collect-profiles-29412975-hfbt9\" (UID: \"f249b9c6-1ad9-44b0-8475-8c11dd212f01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412975-hfbt9" Dec 03 16:15:00 crc kubenswrapper[4998]: I1203 16:15:00.351433 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtgwr\" (UniqueName: \"kubernetes.io/projected/f249b9c6-1ad9-44b0-8475-8c11dd212f01-kube-api-access-gtgwr\") pod \"collect-profiles-29412975-hfbt9\" (UID: \"f249b9c6-1ad9-44b0-8475-8c11dd212f01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412975-hfbt9" Dec 03 16:15:00 crc kubenswrapper[4998]: I1203 16:15:00.489973 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412975-hfbt9" Dec 03 16:15:00 crc kubenswrapper[4998]: I1203 16:15:00.682023 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412975-hfbt9"] Dec 03 16:15:00 crc kubenswrapper[4998]: I1203 16:15:00.755503 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412975-hfbt9" event={"ID":"f249b9c6-1ad9-44b0-8475-8c11dd212f01","Type":"ContainerStarted","Data":"d7b2fe585b563f8c9d6928fb6d9323a867d903419a23e2590cf3c9fd3002e18e"} Dec 03 16:15:01 crc kubenswrapper[4998]: I1203 16:15:01.763612 4998 generic.go:334] "Generic (PLEG): container finished" podID="f249b9c6-1ad9-44b0-8475-8c11dd212f01" containerID="d82d3f9946783abd8003d19ede7c6ae7f58f97e12e99fa71b406eb485cb1eb8e" exitCode=0 Dec 03 16:15:01 crc kubenswrapper[4998]: I1203 16:15:01.763729 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412975-hfbt9" event={"ID":"f249b9c6-1ad9-44b0-8475-8c11dd212f01","Type":"ContainerDied","Data":"d82d3f9946783abd8003d19ede7c6ae7f58f97e12e99fa71b406eb485cb1eb8e"} Dec 03 16:15:03 crc kubenswrapper[4998]: I1203 16:15:03.014772 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412975-hfbt9" Dec 03 16:15:03 crc kubenswrapper[4998]: I1203 16:15:03.066853 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f249b9c6-1ad9-44b0-8475-8c11dd212f01-secret-volume\") pod \"f249b9c6-1ad9-44b0-8475-8c11dd212f01\" (UID: \"f249b9c6-1ad9-44b0-8475-8c11dd212f01\") " Dec 03 16:15:03 crc kubenswrapper[4998]: I1203 16:15:03.066927 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtgwr\" (UniqueName: \"kubernetes.io/projected/f249b9c6-1ad9-44b0-8475-8c11dd212f01-kube-api-access-gtgwr\") pod \"f249b9c6-1ad9-44b0-8475-8c11dd212f01\" (UID: \"f249b9c6-1ad9-44b0-8475-8c11dd212f01\") " Dec 03 16:15:03 crc kubenswrapper[4998]: I1203 16:15:03.066957 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f249b9c6-1ad9-44b0-8475-8c11dd212f01-config-volume\") pod \"f249b9c6-1ad9-44b0-8475-8c11dd212f01\" (UID: \"f249b9c6-1ad9-44b0-8475-8c11dd212f01\") " Dec 03 16:15:03 crc kubenswrapper[4998]: I1203 16:15:03.067800 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f249b9c6-1ad9-44b0-8475-8c11dd212f01-config-volume" (OuterVolumeSpecName: "config-volume") pod "f249b9c6-1ad9-44b0-8475-8c11dd212f01" (UID: "f249b9c6-1ad9-44b0-8475-8c11dd212f01"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:15:03 crc kubenswrapper[4998]: I1203 16:15:03.073294 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f249b9c6-1ad9-44b0-8475-8c11dd212f01-kube-api-access-gtgwr" (OuterVolumeSpecName: "kube-api-access-gtgwr") pod "f249b9c6-1ad9-44b0-8475-8c11dd212f01" (UID: "f249b9c6-1ad9-44b0-8475-8c11dd212f01"). InnerVolumeSpecName "kube-api-access-gtgwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:15:03 crc kubenswrapper[4998]: I1203 16:15:03.075297 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f249b9c6-1ad9-44b0-8475-8c11dd212f01-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f249b9c6-1ad9-44b0-8475-8c11dd212f01" (UID: "f249b9c6-1ad9-44b0-8475-8c11dd212f01"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:15:03 crc kubenswrapper[4998]: I1203 16:15:03.168078 4998 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f249b9c6-1ad9-44b0-8475-8c11dd212f01-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:03 crc kubenswrapper[4998]: I1203 16:15:03.168418 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtgwr\" (UniqueName: \"kubernetes.io/projected/f249b9c6-1ad9-44b0-8475-8c11dd212f01-kube-api-access-gtgwr\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:03 crc kubenswrapper[4998]: I1203 16:15:03.168508 4998 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f249b9c6-1ad9-44b0-8475-8c11dd212f01-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:03 crc kubenswrapper[4998]: I1203 16:15:03.777855 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412975-hfbt9" event={"ID":"f249b9c6-1ad9-44b0-8475-8c11dd212f01","Type":"ContainerDied","Data":"d7b2fe585b563f8c9d6928fb6d9323a867d903419a23e2590cf3c9fd3002e18e"} Dec 03 16:15:03 crc kubenswrapper[4998]: I1203 16:15:03.777890 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7b2fe585b563f8c9d6928fb6d9323a867d903419a23e2590cf3c9fd3002e18e" Dec 03 16:15:03 crc kubenswrapper[4998]: I1203 16:15:03.777892 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412975-hfbt9" Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.337323 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-p7n29"] Dec 03 16:15:04 crc kubenswrapper[4998]: E1203 16:15:04.337554 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f249b9c6-1ad9-44b0-8475-8c11dd212f01" containerName="collect-profiles" Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.337568 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="f249b9c6-1ad9-44b0-8475-8c11dd212f01" containerName="collect-profiles" Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.337672 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="f249b9c6-1ad9-44b0-8475-8c11dd212f01" containerName="collect-profiles" Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.338080 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-p7n29" Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.346295 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.346580 4998 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-db5zl" Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.346627 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.350747 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-7sddv"] Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.351827 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-7sddv" Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.356057 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-p7n29"] Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.359263 4998 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-rg5g6" Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.378601 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-xcmb6"] Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.379482 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-xcmb6" Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.381677 4998 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-rzjjm" Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.382796 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-7sddv"] Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.386185 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2vtp\" (UniqueName: \"kubernetes.io/projected/b12a5054-c73c-4d99-a5a8-174decd7d642-kube-api-access-b2vtp\") pod \"cert-manager-cainjector-7f985d654d-p7n29\" (UID: \"b12a5054-c73c-4d99-a5a8-174decd7d642\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-p7n29" Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.387503 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-xcmb6"] Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.487668 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2rzm\" (UniqueName: \"kubernetes.io/projected/f4bb9920-07e2-4f7d-a81b-e91c374e8f28-kube-api-access-j2rzm\") pod \"cert-manager-webhook-5655c58dd6-xcmb6\" (UID: \"f4bb9920-07e2-4f7d-a81b-e91c374e8f28\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-xcmb6" Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.488094 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2vtp\" (UniqueName: \"kubernetes.io/projected/b12a5054-c73c-4d99-a5a8-174decd7d642-kube-api-access-b2vtp\") pod \"cert-manager-cainjector-7f985d654d-p7n29\" (UID: \"b12a5054-c73c-4d99-a5a8-174decd7d642\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-p7n29" Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.488126 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwjf2\" (UniqueName: \"kubernetes.io/projected/ab951659-919a-4a60-b494-56278d2cd184-kube-api-access-zwjf2\") pod \"cert-manager-5b446d88c5-7sddv\" (UID: \"ab951659-919a-4a60-b494-56278d2cd184\") " pod="cert-manager/cert-manager-5b446d88c5-7sddv" Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.505428 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2vtp\" (UniqueName: \"kubernetes.io/projected/b12a5054-c73c-4d99-a5a8-174decd7d642-kube-api-access-b2vtp\") pod \"cert-manager-cainjector-7f985d654d-p7n29\" (UID: \"b12a5054-c73c-4d99-a5a8-174decd7d642\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-p7n29" Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.589268 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2rzm\" (UniqueName: \"kubernetes.io/projected/f4bb9920-07e2-4f7d-a81b-e91c374e8f28-kube-api-access-j2rzm\") pod \"cert-manager-webhook-5655c58dd6-xcmb6\" (UID: \"f4bb9920-07e2-4f7d-a81b-e91c374e8f28\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-xcmb6" Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.589343 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwjf2\" (UniqueName: \"kubernetes.io/projected/ab951659-919a-4a60-b494-56278d2cd184-kube-api-access-zwjf2\") pod \"cert-manager-5b446d88c5-7sddv\" (UID: \"ab951659-919a-4a60-b494-56278d2cd184\") " pod="cert-manager/cert-manager-5b446d88c5-7sddv" Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.603516 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwjf2\" (UniqueName: \"kubernetes.io/projected/ab951659-919a-4a60-b494-56278d2cd184-kube-api-access-zwjf2\") pod \"cert-manager-5b446d88c5-7sddv\" (UID: \"ab951659-919a-4a60-b494-56278d2cd184\") " pod="cert-manager/cert-manager-5b446d88c5-7sddv" Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.604577 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2rzm\" (UniqueName: \"kubernetes.io/projected/f4bb9920-07e2-4f7d-a81b-e91c374e8f28-kube-api-access-j2rzm\") pod \"cert-manager-webhook-5655c58dd6-xcmb6\" (UID: \"f4bb9920-07e2-4f7d-a81b-e91c374e8f28\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-xcmb6" Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.677186 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-p7n29" Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.685846 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-7sddv" Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.696193 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-xcmb6" Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.953209 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-7sddv"] Dec 03 16:15:04 crc kubenswrapper[4998]: I1203 16:15:04.963717 4998 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 16:15:05 crc kubenswrapper[4998]: I1203 16:15:05.106342 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-p7n29"] Dec 03 16:15:05 crc kubenswrapper[4998]: W1203 16:15:05.108143 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb12a5054_c73c_4d99_a5a8_174decd7d642.slice/crio-8677eb38500b1dbe3aa45675023e982af490997f4b79888a4f68c1fb0559a0b8 WatchSource:0}: Error finding container 8677eb38500b1dbe3aa45675023e982af490997f4b79888a4f68c1fb0559a0b8: Status 404 returned error can't find the container with id 8677eb38500b1dbe3aa45675023e982af490997f4b79888a4f68c1fb0559a0b8 Dec 03 16:15:05 crc kubenswrapper[4998]: I1203 16:15:05.182079 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-xcmb6"] Dec 03 16:15:05 crc kubenswrapper[4998]: W1203 16:15:05.187426 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4bb9920_07e2_4f7d_a81b_e91c374e8f28.slice/crio-2a079774d06f5211c57b05cb6473eb998f841463e36a57af104057990946a9c1 WatchSource:0}: Error finding container 2a079774d06f5211c57b05cb6473eb998f841463e36a57af104057990946a9c1: Status 404 returned error can't find the container with id 2a079774d06f5211c57b05cb6473eb998f841463e36a57af104057990946a9c1 Dec 03 16:15:05 crc kubenswrapper[4998]: I1203 16:15:05.799193 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-xcmb6" event={"ID":"f4bb9920-07e2-4f7d-a81b-e91c374e8f28","Type":"ContainerStarted","Data":"2a079774d06f5211c57b05cb6473eb998f841463e36a57af104057990946a9c1"} Dec 03 16:15:05 crc kubenswrapper[4998]: I1203 16:15:05.801362 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-7sddv" event={"ID":"ab951659-919a-4a60-b494-56278d2cd184","Type":"ContainerStarted","Data":"6def9d1ec50304880168be90e39849bfc4499af9b7912ce51c6ceeecca4986b3"} Dec 03 16:15:05 crc kubenswrapper[4998]: I1203 16:15:05.803161 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-p7n29" event={"ID":"b12a5054-c73c-4d99-a5a8-174decd7d642","Type":"ContainerStarted","Data":"8677eb38500b1dbe3aa45675023e982af490997f4b79888a4f68c1fb0559a0b8"} Dec 03 16:15:08 crc kubenswrapper[4998]: I1203 16:15:08.832459 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-p7n29" event={"ID":"b12a5054-c73c-4d99-a5a8-174decd7d642","Type":"ContainerStarted","Data":"4b32060a02ccbf1041447fc3377f5661b64de969270c77580bb14f6ad30f2a95"} Dec 03 16:15:08 crc kubenswrapper[4998]: I1203 16:15:08.857288 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-p7n29" podStartSLOduration=1.973996093 podStartE2EDuration="4.857239642s" podCreationTimestamp="2025-12-03 16:15:04 +0000 UTC" firstStartedPulling="2025-12-03 16:15:05.111070667 +0000 UTC m=+683.722770890" lastFinishedPulling="2025-12-03 16:15:07.994314216 +0000 UTC m=+686.606014439" observedRunningTime="2025-12-03 16:15:08.850033939 +0000 UTC m=+687.461734172" watchObservedRunningTime="2025-12-03 16:15:08.857239642 +0000 UTC m=+687.468939865" Dec 03 16:15:09 crc kubenswrapper[4998]: I1203 16:15:09.840531 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-xcmb6" event={"ID":"f4bb9920-07e2-4f7d-a81b-e91c374e8f28","Type":"ContainerStarted","Data":"0132901a3dfd98a8235c3664f7b67d2d783367b7389008113c6bee7b3dd3fb5f"} Dec 03 16:15:09 crc kubenswrapper[4998]: I1203 16:15:09.842233 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-xcmb6" Dec 03 16:15:09 crc kubenswrapper[4998]: I1203 16:15:09.845015 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-7sddv" event={"ID":"ab951659-919a-4a60-b494-56278d2cd184","Type":"ContainerStarted","Data":"1113ab26ddb96aa57eb58ea4a24d7ca3776c1e09da014ed68dd876129520a588"} Dec 03 16:15:09 crc kubenswrapper[4998]: I1203 16:15:09.865524 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-xcmb6" podStartSLOduration=1.425770848 podStartE2EDuration="5.86550074s" podCreationTimestamp="2025-12-03 16:15:04 +0000 UTC" firstStartedPulling="2025-12-03 16:15:05.191051163 +0000 UTC m=+683.802751396" lastFinishedPulling="2025-12-03 16:15:09.630781025 +0000 UTC m=+688.242481288" observedRunningTime="2025-12-03 16:15:09.864369531 +0000 UTC m=+688.476069754" watchObservedRunningTime="2025-12-03 16:15:09.86550074 +0000 UTC m=+688.477200993" Dec 03 16:15:09 crc kubenswrapper[4998]: I1203 16:15:09.884887 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-7sddv" podStartSLOduration=1.277765531 podStartE2EDuration="5.884867153s" podCreationTimestamp="2025-12-03 16:15:04 +0000 UTC" firstStartedPulling="2025-12-03 16:15:04.963520791 +0000 UTC m=+683.575221014" lastFinishedPulling="2025-12-03 16:15:09.570622423 +0000 UTC m=+688.182322636" observedRunningTime="2025-12-03 16:15:09.882528983 +0000 UTC m=+688.494229206" watchObservedRunningTime="2025-12-03 16:15:09.884867153 +0000 UTC m=+688.496567366" Dec 03 16:15:14 crc kubenswrapper[4998]: I1203 16:15:14.698344 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-xcmb6" Dec 03 16:15:14 crc kubenswrapper[4998]: I1203 16:15:14.956540 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-m2lh2"] Dec 03 16:15:14 crc kubenswrapper[4998]: I1203 16:15:14.957030 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="ovn-controller" containerID="cri-o://aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894" gracePeriod=30 Dec 03 16:15:14 crc kubenswrapper[4998]: I1203 16:15:14.957101 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="sbdb" containerID="cri-o://3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538" gracePeriod=30 Dec 03 16:15:14 crc kubenswrapper[4998]: I1203 16:15:14.957134 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246" gracePeriod=30 Dec 03 16:15:14 crc kubenswrapper[4998]: I1203 16:15:14.957175 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="kube-rbac-proxy-node" containerID="cri-o://ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335" gracePeriod=30 Dec 03 16:15:14 crc kubenswrapper[4998]: I1203 16:15:14.957272 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="nbdb" containerID="cri-o://8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b" gracePeriod=30 Dec 03 16:15:14 crc kubenswrapper[4998]: I1203 16:15:14.957347 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="northd" containerID="cri-o://91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423" gracePeriod=30 Dec 03 16:15:14 crc kubenswrapper[4998]: I1203 16:15:14.957387 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="ovn-acl-logging" containerID="cri-o://65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3" gracePeriod=30 Dec 03 16:15:14 crc kubenswrapper[4998]: I1203 16:15:14.997058 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="ovnkube-controller" containerID="cri-o://5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df" gracePeriod=30 Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.239536 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m2lh2_4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9/ovnkube-controller/2.log" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.243723 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m2lh2_4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9/ovn-acl-logging/0.log" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.244540 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m2lh2_4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9/ovn-controller/0.log" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.245153 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.296896 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dtvp7"] Dec 03 16:15:15 crc kubenswrapper[4998]: E1203 16:15:15.297090 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.297102 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 16:15:15 crc kubenswrapper[4998]: E1203 16:15:15.297111 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="nbdb" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.297118 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="nbdb" Dec 03 16:15:15 crc kubenswrapper[4998]: E1203 16:15:15.297128 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="sbdb" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.297134 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="sbdb" Dec 03 16:15:15 crc kubenswrapper[4998]: E1203 16:15:15.297140 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="ovnkube-controller" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.297146 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="ovnkube-controller" Dec 03 16:15:15 crc kubenswrapper[4998]: E1203 16:15:15.297155 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="northd" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.297160 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="northd" Dec 03 16:15:15 crc kubenswrapper[4998]: E1203 16:15:15.297168 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="ovn-controller" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.297175 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="ovn-controller" Dec 03 16:15:15 crc kubenswrapper[4998]: E1203 16:15:15.297181 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="ovnkube-controller" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.297188 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="ovnkube-controller" Dec 03 16:15:15 crc kubenswrapper[4998]: E1203 16:15:15.297195 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="ovnkube-controller" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.297201 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="ovnkube-controller" Dec 03 16:15:15 crc kubenswrapper[4998]: E1203 16:15:15.297208 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="kubecfg-setup" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.297215 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="kubecfg-setup" Dec 03 16:15:15 crc kubenswrapper[4998]: E1203 16:15:15.297223 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="ovn-acl-logging" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.297229 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="ovn-acl-logging" Dec 03 16:15:15 crc kubenswrapper[4998]: E1203 16:15:15.297239 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="kube-rbac-proxy-node" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.297247 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="kube-rbac-proxy-node" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.297358 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="kube-rbac-proxy-node" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.297369 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="ovnkube-controller" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.297379 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="ovn-acl-logging" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.297392 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="northd" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.297401 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="sbdb" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.297411 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.297419 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="ovnkube-controller" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.297428 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="ovn-controller" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.297438 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="nbdb" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.297447 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="ovnkube-controller" Dec 03 16:15:15 crc kubenswrapper[4998]: E1203 16:15:15.297557 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="ovnkube-controller" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.297569 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="ovnkube-controller" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.297679 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerName="ovnkube-controller" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.299597 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.327842 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-cni-bin\") pod \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.327881 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-run-systemd\") pod \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.327901 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-cni-netd\") pod \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.327936 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-ovn-node-metrics-cert\") pod \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.327971 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-ovnkube-config\") pod \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328014 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-slash\") pod \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328030 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-run-ovn-kubernetes\") pod \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328057 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-kubelet\") pod \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328081 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-run-netns\") pod \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328112 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-run-openvswitch\") pod \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328153 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-log-socket\") pod \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328173 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-systemd-units\") pod \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328194 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-ovnkube-script-lib\") pod \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328219 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-run-ovn\") pod \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328237 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328260 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-node-log\") pod \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328275 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-env-overrides\") pod \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328296 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-var-lib-openvswitch\") pod \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328320 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54d7l\" (UniqueName: \"kubernetes.io/projected/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-kube-api-access-54d7l\") pod \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328336 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-etc-openvswitch\") pod \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\" (UID: \"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9\") " Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328382 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" (UID: "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328455 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" (UID: "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328520 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" (UID: "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328541 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" (UID: "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328556 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-log-socket" (OuterVolumeSpecName: "log-socket") pod "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" (UID: "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328570 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" (UID: "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328806 4998 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328847 4998 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328866 4998 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328886 4998 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328905 4998 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-log-socket\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.328924 4998 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.329052 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" (UID: "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.329315 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" (UID: "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.329342 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" (UID: "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.329725 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" (UID: "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.329830 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" (UID: "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.329868 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-node-log" (OuterVolumeSpecName: "node-log") pod "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" (UID: "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.330106 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" (UID: "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.330183 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-slash" (OuterVolumeSpecName: "host-slash") pod "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" (UID: "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.330245 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" (UID: "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.330295 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" (UID: "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.330339 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" (UID: "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.335293 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" (UID: "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.335517 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-kube-api-access-54d7l" (OuterVolumeSpecName: "kube-api-access-54d7l") pod "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" (UID: "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9"). InnerVolumeSpecName "kube-api-access-54d7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.345143 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" (UID: "4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.430294 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-host-run-ovn-kubernetes\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.430385 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-host-cni-netd\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.430420 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-host-run-netns\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.430476 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-host-slash\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.430547 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-systemd-units\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.430583 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-host-kubelet\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.430618 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dc50747a-0a5c-48c3-98d9-f623c3aa2257-ovn-node-metrics-cert\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.430659 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-run-systemd\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.430688 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-log-socket\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.430719 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvjp2\" (UniqueName: \"kubernetes.io/projected/dc50747a-0a5c-48c3-98d9-f623c3aa2257-kube-api-access-wvjp2\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.430745 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dc50747a-0a5c-48c3-98d9-f623c3aa2257-env-overrides\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.430808 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.430854 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-etc-openvswitch\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.430883 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-run-openvswitch\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.430915 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-host-cni-bin\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.430948 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-node-log\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.430977 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dc50747a-0a5c-48c3-98d9-f623c3aa2257-ovnkube-config\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.431062 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/dc50747a-0a5c-48c3-98d9-f623c3aa2257-ovnkube-script-lib\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.431153 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-var-lib-openvswitch\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.431203 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-run-ovn\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.431298 4998 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.431321 4998 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.431336 4998 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.431348 4998 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-node-log\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.431361 4998 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.431373 4998 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.431386 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54d7l\" (UniqueName: \"kubernetes.io/projected/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-kube-api-access-54d7l\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.431397 4998 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.431407 4998 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.431417 4998 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.431428 4998 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.431439 4998 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-slash\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.431452 4998 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.431463 4998 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.532614 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-systemd-units\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.532663 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-host-kubelet\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.532696 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dc50747a-0a5c-48c3-98d9-f623c3aa2257-ovn-node-metrics-cert\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.532719 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-log-socket\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.532814 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-run-systemd\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.532847 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dc50747a-0a5c-48c3-98d9-f623c3aa2257-env-overrides\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.532871 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvjp2\" (UniqueName: \"kubernetes.io/projected/dc50747a-0a5c-48c3-98d9-f623c3aa2257-kube-api-access-wvjp2\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.532901 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.532944 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-etc-openvswitch\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.532969 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-run-openvswitch\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.532992 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-host-cni-bin\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.532987 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-host-kubelet\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.533022 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-node-log\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.533073 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-node-log\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.533129 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-log-socket\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.533133 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dc50747a-0a5c-48c3-98d9-f623c3aa2257-ovnkube-config\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.533162 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-run-systemd\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.533243 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/dc50747a-0a5c-48c3-98d9-f623c3aa2257-ovnkube-script-lib\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.533290 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-etc-openvswitch\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.533346 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-var-lib-openvswitch\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.533427 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.533442 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-run-ovn\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.533546 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-host-run-ovn-kubernetes\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.533590 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-host-cni-netd\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.533615 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-host-run-netns\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.533641 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-host-slash\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.533738 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-host-run-ovn-kubernetes\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.533786 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-host-slash\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.533483 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-run-ovn\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.533807 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dc50747a-0a5c-48c3-98d9-f623c3aa2257-env-overrides\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.533829 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-host-cni-netd\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.533248 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-systemd-units\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.533859 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-host-cni-bin\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.533864 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-var-lib-openvswitch\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.533827 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-run-openvswitch\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.533878 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dc50747a-0a5c-48c3-98d9-f623c3aa2257-host-run-netns\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.533986 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dc50747a-0a5c-48c3-98d9-f623c3aa2257-ovnkube-config\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.534919 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/dc50747a-0a5c-48c3-98d9-f623c3aa2257-ovnkube-script-lib\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.535864 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dc50747a-0a5c-48c3-98d9-f623c3aa2257-ovn-node-metrics-cert\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.550243 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvjp2\" (UniqueName: \"kubernetes.io/projected/dc50747a-0a5c-48c3-98d9-f623c3aa2257-kube-api-access-wvjp2\") pod \"ovnkube-node-dtvp7\" (UID: \"dc50747a-0a5c-48c3-98d9-f623c3aa2257\") " pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.614704 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:15 crc kubenswrapper[4998]: W1203 16:15:15.645798 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc50747a_0a5c_48c3_98d9_f623c3aa2257.slice/crio-161e46de4bbe77bb5c0e694617f474085cc9840b14a7ec318548f27b414424c4 WatchSource:0}: Error finding container 161e46de4bbe77bb5c0e694617f474085cc9840b14a7ec318548f27b414424c4: Status 404 returned error can't find the container with id 161e46de4bbe77bb5c0e694617f474085cc9840b14a7ec318548f27b414424c4 Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.891885 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m2lh2_4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9/ovnkube-controller/2.log" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.894538 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m2lh2_4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9/ovn-acl-logging/0.log" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.895418 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-m2lh2_4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9/ovn-controller/0.log" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.895876 4998 generic.go:334] "Generic (PLEG): container finished" podID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerID="5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df" exitCode=0 Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.895922 4998 generic.go:334] "Generic (PLEG): container finished" podID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerID="3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538" exitCode=0 Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.895934 4998 generic.go:334] "Generic (PLEG): container finished" podID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerID="8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b" exitCode=0 Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.895943 4998 generic.go:334] "Generic (PLEG): container finished" podID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerID="91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423" exitCode=0 Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.895950 4998 generic.go:334] "Generic (PLEG): container finished" podID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerID="1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246" exitCode=0 Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.895957 4998 generic.go:334] "Generic (PLEG): container finished" podID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerID="ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335" exitCode=0 Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.895966 4998 generic.go:334] "Generic (PLEG): container finished" podID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerID="65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3" exitCode=143 Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.895972 4998 generic.go:334] "Generic (PLEG): container finished" podID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" containerID="aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894" exitCode=143 Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896030 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" event={"ID":"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9","Type":"ContainerDied","Data":"5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896114 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" event={"ID":"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9","Type":"ContainerDied","Data":"3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896149 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" event={"ID":"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9","Type":"ContainerDied","Data":"8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896173 4998 scope.go:117] "RemoveContainer" containerID="5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896199 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896176 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" event={"ID":"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9","Type":"ContainerDied","Data":"91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896337 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" event={"ID":"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9","Type":"ContainerDied","Data":"1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896369 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" event={"ID":"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9","Type":"ContainerDied","Data":"ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896395 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896419 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896437 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896453 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896467 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896482 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896496 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896512 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896526 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896560 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" event={"ID":"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9","Type":"ContainerDied","Data":"65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896584 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896603 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896619 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896634 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896648 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896663 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896678 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896692 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896707 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896722 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896741 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" event={"ID":"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9","Type":"ContainerDied","Data":"aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896807 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896827 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896841 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896855 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896868 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896883 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896898 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896912 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896926 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896939 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896958 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-m2lh2" event={"ID":"4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9","Type":"ContainerDied","Data":"c740368518ed3aa02e3b48f5e955778758c5b3ba693eef7207af578dd11186ee"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.896982 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.897000 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.897013 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.897124 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.897144 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.897158 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.897173 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.897187 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.897202 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.897217 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.899531 4998 generic.go:334] "Generic (PLEG): container finished" podID="dc50747a-0a5c-48c3-98d9-f623c3aa2257" containerID="eb06a17a8d9661827feb708fe5d9f3d64dd858ec655c01594764a993a4620690" exitCode=0 Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.899626 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" event={"ID":"dc50747a-0a5c-48c3-98d9-f623c3aa2257","Type":"ContainerDied","Data":"eb06a17a8d9661827feb708fe5d9f3d64dd858ec655c01594764a993a4620690"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.899660 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" event={"ID":"dc50747a-0a5c-48c3-98d9-f623c3aa2257","Type":"ContainerStarted","Data":"161e46de4bbe77bb5c0e694617f474085cc9840b14a7ec318548f27b414424c4"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.903617 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fcdxh_bd9d66fb-a400-4810-aa7a-c81c9c24bd11/kube-multus/1.log" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.905375 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fcdxh_bd9d66fb-a400-4810-aa7a-c81c9c24bd11/kube-multus/0.log" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.905456 4998 generic.go:334] "Generic (PLEG): container finished" podID="bd9d66fb-a400-4810-aa7a-c81c9c24bd11" containerID="1341e3406e28b4e4309aaa18fba25c82ff489cc9cb44f232c4eb0f9a3762205c" exitCode=2 Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.905519 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fcdxh" event={"ID":"bd9d66fb-a400-4810-aa7a-c81c9c24bd11","Type":"ContainerDied","Data":"1341e3406e28b4e4309aaa18fba25c82ff489cc9cb44f232c4eb0f9a3762205c"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.905550 4998 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d"} Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.906875 4998 scope.go:117] "RemoveContainer" containerID="1341e3406e28b4e4309aaa18fba25c82ff489cc9cb44f232c4eb0f9a3762205c" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.925001 4998 scope.go:117] "RemoveContainer" containerID="05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958" Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.958539 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-m2lh2"] Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.962641 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-m2lh2"] Dec 03 16:15:15 crc kubenswrapper[4998]: I1203 16:15:15.975984 4998 scope.go:117] "RemoveContainer" containerID="3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.009790 4998 scope.go:117] "RemoveContainer" containerID="8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.025310 4998 scope.go:117] "RemoveContainer" containerID="91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.045475 4998 scope.go:117] "RemoveContainer" containerID="1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.057984 4998 scope.go:117] "RemoveContainer" containerID="ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.076487 4998 scope.go:117] "RemoveContainer" containerID="65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.107738 4998 scope.go:117] "RemoveContainer" containerID="aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.133445 4998 scope.go:117] "RemoveContainer" containerID="eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.153922 4998 scope.go:117] "RemoveContainer" containerID="5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df" Dec 03 16:15:16 crc kubenswrapper[4998]: E1203 16:15:16.155856 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df\": container with ID starting with 5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df not found: ID does not exist" containerID="5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.155921 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df"} err="failed to get container status \"5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df\": rpc error: code = NotFound desc = could not find container \"5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df\": container with ID starting with 5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.155948 4998 scope.go:117] "RemoveContainer" containerID="05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958" Dec 03 16:15:16 crc kubenswrapper[4998]: E1203 16:15:16.157313 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958\": container with ID starting with 05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958 not found: ID does not exist" containerID="05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.157454 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958"} err="failed to get container status \"05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958\": rpc error: code = NotFound desc = could not find container \"05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958\": container with ID starting with 05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.157803 4998 scope.go:117] "RemoveContainer" containerID="3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538" Dec 03 16:15:16 crc kubenswrapper[4998]: E1203 16:15:16.159291 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\": container with ID starting with 3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538 not found: ID does not exist" containerID="3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.160048 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538"} err="failed to get container status \"3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\": rpc error: code = NotFound desc = could not find container \"3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\": container with ID starting with 3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.160111 4998 scope.go:117] "RemoveContainer" containerID="8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b" Dec 03 16:15:16 crc kubenswrapper[4998]: E1203 16:15:16.160803 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\": container with ID starting with 8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b not found: ID does not exist" containerID="8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.160839 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b"} err="failed to get container status \"8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\": rpc error: code = NotFound desc = could not find container \"8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\": container with ID starting with 8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.160865 4998 scope.go:117] "RemoveContainer" containerID="91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423" Dec 03 16:15:16 crc kubenswrapper[4998]: E1203 16:15:16.161646 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\": container with ID starting with 91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423 not found: ID does not exist" containerID="91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.161673 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423"} err="failed to get container status \"91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\": rpc error: code = NotFound desc = could not find container \"91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\": container with ID starting with 91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.161688 4998 scope.go:117] "RemoveContainer" containerID="1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246" Dec 03 16:15:16 crc kubenswrapper[4998]: E1203 16:15:16.162062 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\": container with ID starting with 1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246 not found: ID does not exist" containerID="1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.162079 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246"} err="failed to get container status \"1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\": rpc error: code = NotFound desc = could not find container \"1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\": container with ID starting with 1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.162091 4998 scope.go:117] "RemoveContainer" containerID="ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335" Dec 03 16:15:16 crc kubenswrapper[4998]: E1203 16:15:16.162708 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\": container with ID starting with ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335 not found: ID does not exist" containerID="ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.162729 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335"} err="failed to get container status \"ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\": rpc error: code = NotFound desc = could not find container \"ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\": container with ID starting with ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.162741 4998 scope.go:117] "RemoveContainer" containerID="65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3" Dec 03 16:15:16 crc kubenswrapper[4998]: E1203 16:15:16.163628 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\": container with ID starting with 65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3 not found: ID does not exist" containerID="65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.165034 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3"} err="failed to get container status \"65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\": rpc error: code = NotFound desc = could not find container \"65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\": container with ID starting with 65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.165091 4998 scope.go:117] "RemoveContainer" containerID="aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894" Dec 03 16:15:16 crc kubenswrapper[4998]: E1203 16:15:16.165459 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\": container with ID starting with aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894 not found: ID does not exist" containerID="aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.165484 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894"} err="failed to get container status \"aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\": rpc error: code = NotFound desc = could not find container \"aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\": container with ID starting with aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.165504 4998 scope.go:117] "RemoveContainer" containerID="eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8" Dec 03 16:15:16 crc kubenswrapper[4998]: E1203 16:15:16.165877 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\": container with ID starting with eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8 not found: ID does not exist" containerID="eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.165898 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8"} err="failed to get container status \"eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\": rpc error: code = NotFound desc = could not find container \"eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\": container with ID starting with eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.165916 4998 scope.go:117] "RemoveContainer" containerID="5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.166829 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df"} err="failed to get container status \"5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df\": rpc error: code = NotFound desc = could not find container \"5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df\": container with ID starting with 5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.166866 4998 scope.go:117] "RemoveContainer" containerID="05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.167351 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958"} err="failed to get container status \"05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958\": rpc error: code = NotFound desc = could not find container \"05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958\": container with ID starting with 05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.167401 4998 scope.go:117] "RemoveContainer" containerID="3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.167877 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538"} err="failed to get container status \"3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\": rpc error: code = NotFound desc = could not find container \"3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\": container with ID starting with 3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.167916 4998 scope.go:117] "RemoveContainer" containerID="8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.168683 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b"} err="failed to get container status \"8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\": rpc error: code = NotFound desc = could not find container \"8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\": container with ID starting with 8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.168726 4998 scope.go:117] "RemoveContainer" containerID="91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.169284 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423"} err="failed to get container status \"91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\": rpc error: code = NotFound desc = could not find container \"91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\": container with ID starting with 91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.169325 4998 scope.go:117] "RemoveContainer" containerID="1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.169856 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246"} err="failed to get container status \"1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\": rpc error: code = NotFound desc = could not find container \"1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\": container with ID starting with 1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.169878 4998 scope.go:117] "RemoveContainer" containerID="ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.170281 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335"} err="failed to get container status \"ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\": rpc error: code = NotFound desc = could not find container \"ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\": container with ID starting with ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.170316 4998 scope.go:117] "RemoveContainer" containerID="65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.170543 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3"} err="failed to get container status \"65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\": rpc error: code = NotFound desc = could not find container \"65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\": container with ID starting with 65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.170563 4998 scope.go:117] "RemoveContainer" containerID="aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.170896 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894"} err="failed to get container status \"aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\": rpc error: code = NotFound desc = could not find container \"aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\": container with ID starting with aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.170933 4998 scope.go:117] "RemoveContainer" containerID="eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.171421 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8"} err="failed to get container status \"eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\": rpc error: code = NotFound desc = could not find container \"eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\": container with ID starting with eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.171445 4998 scope.go:117] "RemoveContainer" containerID="5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.171747 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df"} err="failed to get container status \"5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df\": rpc error: code = NotFound desc = could not find container \"5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df\": container with ID starting with 5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.171803 4998 scope.go:117] "RemoveContainer" containerID="05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.172171 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958"} err="failed to get container status \"05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958\": rpc error: code = NotFound desc = could not find container \"05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958\": container with ID starting with 05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.172196 4998 scope.go:117] "RemoveContainer" containerID="3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.172435 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538"} err="failed to get container status \"3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\": rpc error: code = NotFound desc = could not find container \"3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\": container with ID starting with 3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.172455 4998 scope.go:117] "RemoveContainer" containerID="8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.172762 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b"} err="failed to get container status \"8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\": rpc error: code = NotFound desc = could not find container \"8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\": container with ID starting with 8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.172790 4998 scope.go:117] "RemoveContainer" containerID="91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.173074 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423"} err="failed to get container status \"91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\": rpc error: code = NotFound desc = could not find container \"91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\": container with ID starting with 91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.173094 4998 scope.go:117] "RemoveContainer" containerID="1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.173400 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246"} err="failed to get container status \"1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\": rpc error: code = NotFound desc = could not find container \"1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\": container with ID starting with 1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.173418 4998 scope.go:117] "RemoveContainer" containerID="ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.173663 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335"} err="failed to get container status \"ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\": rpc error: code = NotFound desc = could not find container \"ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\": container with ID starting with ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.173693 4998 scope.go:117] "RemoveContainer" containerID="65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.173971 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3"} err="failed to get container status \"65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\": rpc error: code = NotFound desc = could not find container \"65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\": container with ID starting with 65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.173998 4998 scope.go:117] "RemoveContainer" containerID="aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.174274 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894"} err="failed to get container status \"aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\": rpc error: code = NotFound desc = could not find container \"aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\": container with ID starting with aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.174291 4998 scope.go:117] "RemoveContainer" containerID="eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.174612 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8"} err="failed to get container status \"eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\": rpc error: code = NotFound desc = could not find container \"eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\": container with ID starting with eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.174630 4998 scope.go:117] "RemoveContainer" containerID="5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.174832 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df"} err="failed to get container status \"5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df\": rpc error: code = NotFound desc = could not find container \"5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df\": container with ID starting with 5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.174861 4998 scope.go:117] "RemoveContainer" containerID="05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.175033 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958"} err="failed to get container status \"05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958\": rpc error: code = NotFound desc = could not find container \"05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958\": container with ID starting with 05f579ee5784e05b096c0e6b25be6a26fb65f5fcd7b36e301d19581cd594f958 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.175045 4998 scope.go:117] "RemoveContainer" containerID="3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.176017 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538"} err="failed to get container status \"3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\": rpc error: code = NotFound desc = could not find container \"3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538\": container with ID starting with 3c6431afc833b06aca9f49f900f00c211a162b76bf2dbcfbc2ef89a17dc69538 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.176034 4998 scope.go:117] "RemoveContainer" containerID="8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.176367 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b"} err="failed to get container status \"8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\": rpc error: code = NotFound desc = could not find container \"8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b\": container with ID starting with 8f0cceeadf3be08be2f9e6726bf2ec8e3deaa379c20821d810ee021c2907158b not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.176383 4998 scope.go:117] "RemoveContainer" containerID="91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.176627 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423"} err="failed to get container status \"91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\": rpc error: code = NotFound desc = could not find container \"91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423\": container with ID starting with 91c4d2abb56a5913dcdcf7da3bbd6baa6ba65148adda7d06d8f2e10817b81423 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.176641 4998 scope.go:117] "RemoveContainer" containerID="1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.176934 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246"} err="failed to get container status \"1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\": rpc error: code = NotFound desc = could not find container \"1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246\": container with ID starting with 1f9106c42ebdd3a24bbc0cd0588e5c885dfdbcbb08e8ae5fb3b62727b0f2a246 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.176947 4998 scope.go:117] "RemoveContainer" containerID="ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.177250 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335"} err="failed to get container status \"ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\": rpc error: code = NotFound desc = could not find container \"ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335\": container with ID starting with ff555a76c590d7ffb0838f5b996862f55706a60bdd5684b1b2127fa12208d335 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.177272 4998 scope.go:117] "RemoveContainer" containerID="65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.177552 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3"} err="failed to get container status \"65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\": rpc error: code = NotFound desc = could not find container \"65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3\": container with ID starting with 65d1ef4dc47497990e2fd23ba8b34725432db4885fefc06df74f9d0cc0fafca3 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.177566 4998 scope.go:117] "RemoveContainer" containerID="aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.177871 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894"} err="failed to get container status \"aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\": rpc error: code = NotFound desc = could not find container \"aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894\": container with ID starting with aad26a7bf6398e81aa3cd8cb6489dc138bc460699d90bd3fa0b7ae8773dc0894 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.177889 4998 scope.go:117] "RemoveContainer" containerID="eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.178366 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8"} err="failed to get container status \"eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\": rpc error: code = NotFound desc = could not find container \"eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8\": container with ID starting with eb8c9e116aa4276c70b1c24375568418c55568f9c3a9ace2525bb4c1a0c8add8 not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.178387 4998 scope.go:117] "RemoveContainer" containerID="5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.178717 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df"} err="failed to get container status \"5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df\": rpc error: code = NotFound desc = could not find container \"5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df\": container with ID starting with 5ac9985739a48d6367702987405aa4e024ce395a158f0675545a3801391e31df not found: ID does not exist" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.922208 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" event={"ID":"dc50747a-0a5c-48c3-98d9-f623c3aa2257","Type":"ContainerStarted","Data":"2a8fdd470691ef725868fa2a51a40c6df8e9bf69f5a23d27231f9c030ab5e807"} Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.922841 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" event={"ID":"dc50747a-0a5c-48c3-98d9-f623c3aa2257","Type":"ContainerStarted","Data":"32a286303c05f00d468790b3009901df96cf3e08ef06dd4d2a309ecbdf7f9b7f"} Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.922920 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" event={"ID":"dc50747a-0a5c-48c3-98d9-f623c3aa2257","Type":"ContainerStarted","Data":"4536349af958ad4a3a3a8cc3b7719abe5344a9a6359e4770c5335c8a2b15a7ae"} Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.922948 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" event={"ID":"dc50747a-0a5c-48c3-98d9-f623c3aa2257","Type":"ContainerStarted","Data":"18d651671ef87416804dd3275c6b721531b37f69be16d042943b5a4cba87db93"} Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.923006 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" event={"ID":"dc50747a-0a5c-48c3-98d9-f623c3aa2257","Type":"ContainerStarted","Data":"bad481ef99b4bc2bac12d07c61355bb1baf4c3b7b8769e77fdd94288877d2799"} Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.923029 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" event={"ID":"dc50747a-0a5c-48c3-98d9-f623c3aa2257","Type":"ContainerStarted","Data":"aac537e9e3ccb87b075c2f0259db20fccc1e33d50775b618d2aba491f8fa1565"} Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.926102 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fcdxh_bd9d66fb-a400-4810-aa7a-c81c9c24bd11/kube-multus/1.log" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.927249 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fcdxh_bd9d66fb-a400-4810-aa7a-c81c9c24bd11/kube-multus/0.log" Dec 03 16:15:16 crc kubenswrapper[4998]: I1203 16:15:16.927311 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fcdxh" event={"ID":"bd9d66fb-a400-4810-aa7a-c81c9c24bd11","Type":"ContainerStarted","Data":"6ad47275e2646c8a2c1311220aeec45a85a8c64f2b7db4bd2a251165b0484e35"} Dec 03 16:15:17 crc kubenswrapper[4998]: I1203 16:15:17.690924 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9" path="/var/lib/kubelet/pods/4ab5b995-c89f-4ea3-9ab2-a0ffa64c64e9/volumes" Dec 03 16:15:18 crc kubenswrapper[4998]: I1203 16:15:18.950504 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" event={"ID":"dc50747a-0a5c-48c3-98d9-f623c3aa2257","Type":"ContainerStarted","Data":"2975e73c58a122129c644e5352899f5c7a7cd45685ed8f699888ec90bbc37211"} Dec 03 16:15:21 crc kubenswrapper[4998]: I1203 16:15:21.977032 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" event={"ID":"dc50747a-0a5c-48c3-98d9-f623c3aa2257","Type":"ContainerStarted","Data":"fb7f199e53890fd5783f409ed7c0a6d584195b452f0405014d7afb82018a2a12"} Dec 03 16:15:21 crc kubenswrapper[4998]: I1203 16:15:21.977525 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:21 crc kubenswrapper[4998]: I1203 16:15:21.977554 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:22 crc kubenswrapper[4998]: I1203 16:15:22.069070 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" podStartSLOduration=7.068734045 podStartE2EDuration="7.068734045s" podCreationTimestamp="2025-12-03 16:15:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:15:22.063767759 +0000 UTC m=+700.675467982" watchObservedRunningTime="2025-12-03 16:15:22.068734045 +0000 UTC m=+700.680434278" Dec 03 16:15:22 crc kubenswrapper[4998]: I1203 16:15:22.089164 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:22 crc kubenswrapper[4998]: I1203 16:15:22.983788 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:23 crc kubenswrapper[4998]: I1203 16:15:23.022320 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:25 crc kubenswrapper[4998]: I1203 16:15:25.273617 4998 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 16:15:27 crc kubenswrapper[4998]: I1203 16:15:27.111808 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:15:27 crc kubenswrapper[4998]: I1203 16:15:27.111914 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:15:41 crc kubenswrapper[4998]: I1203 16:15:41.965958 4998 scope.go:117] "RemoveContainer" containerID="d883a8c4abe0a1a5ab66dc0b9e350a884004f6788c6624be596c49bad15b7b3d" Dec 03 16:15:42 crc kubenswrapper[4998]: I1203 16:15:42.119203 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fcdxh_bd9d66fb-a400-4810-aa7a-c81c9c24bd11/kube-multus/1.log" Dec 03 16:15:43 crc kubenswrapper[4998]: I1203 16:15:43.213435 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj"] Dec 03 16:15:43 crc kubenswrapper[4998]: I1203 16:15:43.215167 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj" Dec 03 16:15:43 crc kubenswrapper[4998]: I1203 16:15:43.219378 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 16:15:43 crc kubenswrapper[4998]: I1203 16:15:43.228316 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj"] Dec 03 16:15:43 crc kubenswrapper[4998]: I1203 16:15:43.300404 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdt2f\" (UniqueName: \"kubernetes.io/projected/9038bb91-6c7d-42c6-b864-9d522c73c488-kube-api-access-cdt2f\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj\" (UID: \"9038bb91-6c7d-42c6-b864-9d522c73c488\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj" Dec 03 16:15:43 crc kubenswrapper[4998]: I1203 16:15:43.300453 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9038bb91-6c7d-42c6-b864-9d522c73c488-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj\" (UID: \"9038bb91-6c7d-42c6-b864-9d522c73c488\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj" Dec 03 16:15:43 crc kubenswrapper[4998]: I1203 16:15:43.300544 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9038bb91-6c7d-42c6-b864-9d522c73c488-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj\" (UID: \"9038bb91-6c7d-42c6-b864-9d522c73c488\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj" Dec 03 16:15:43 crc kubenswrapper[4998]: I1203 16:15:43.401861 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdt2f\" (UniqueName: \"kubernetes.io/projected/9038bb91-6c7d-42c6-b864-9d522c73c488-kube-api-access-cdt2f\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj\" (UID: \"9038bb91-6c7d-42c6-b864-9d522c73c488\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj" Dec 03 16:15:43 crc kubenswrapper[4998]: I1203 16:15:43.401930 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9038bb91-6c7d-42c6-b864-9d522c73c488-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj\" (UID: \"9038bb91-6c7d-42c6-b864-9d522c73c488\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj" Dec 03 16:15:43 crc kubenswrapper[4998]: I1203 16:15:43.402036 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9038bb91-6c7d-42c6-b864-9d522c73c488-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj\" (UID: \"9038bb91-6c7d-42c6-b864-9d522c73c488\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj" Dec 03 16:15:43 crc kubenswrapper[4998]: I1203 16:15:43.403042 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9038bb91-6c7d-42c6-b864-9d522c73c488-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj\" (UID: \"9038bb91-6c7d-42c6-b864-9d522c73c488\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj" Dec 03 16:15:43 crc kubenswrapper[4998]: I1203 16:15:43.403588 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9038bb91-6c7d-42c6-b864-9d522c73c488-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj\" (UID: \"9038bb91-6c7d-42c6-b864-9d522c73c488\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj" Dec 03 16:15:43 crc kubenswrapper[4998]: I1203 16:15:43.426598 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdt2f\" (UniqueName: \"kubernetes.io/projected/9038bb91-6c7d-42c6-b864-9d522c73c488-kube-api-access-cdt2f\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj\" (UID: \"9038bb91-6c7d-42c6-b864-9d522c73c488\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj" Dec 03 16:15:43 crc kubenswrapper[4998]: I1203 16:15:43.534244 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj" Dec 03 16:15:43 crc kubenswrapper[4998]: I1203 16:15:43.756342 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj"] Dec 03 16:15:43 crc kubenswrapper[4998]: W1203 16:15:43.766491 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9038bb91_6c7d_42c6_b864_9d522c73c488.slice/crio-bcd420968c3b8bf20ecb3c9c328563e9149f89baee492ef2ad9cf30fb1666209 WatchSource:0}: Error finding container bcd420968c3b8bf20ecb3c9c328563e9149f89baee492ef2ad9cf30fb1666209: Status 404 returned error can't find the container with id bcd420968c3b8bf20ecb3c9c328563e9149f89baee492ef2ad9cf30fb1666209 Dec 03 16:15:44 crc kubenswrapper[4998]: I1203 16:15:44.136236 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj" event={"ID":"9038bb91-6c7d-42c6-b864-9d522c73c488","Type":"ContainerStarted","Data":"a209638b76d4ea7161615ce8fafb3ab2b6292329d6043441f11c92a086f3285a"} Dec 03 16:15:44 crc kubenswrapper[4998]: I1203 16:15:44.136294 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj" event={"ID":"9038bb91-6c7d-42c6-b864-9d522c73c488","Type":"ContainerStarted","Data":"bcd420968c3b8bf20ecb3c9c328563e9149f89baee492ef2ad9cf30fb1666209"} Dec 03 16:15:45 crc kubenswrapper[4998]: I1203 16:15:45.143422 4998 generic.go:334] "Generic (PLEG): container finished" podID="9038bb91-6c7d-42c6-b864-9d522c73c488" containerID="a209638b76d4ea7161615ce8fafb3ab2b6292329d6043441f11c92a086f3285a" exitCode=0 Dec 03 16:15:45 crc kubenswrapper[4998]: I1203 16:15:45.143539 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj" event={"ID":"9038bb91-6c7d-42c6-b864-9d522c73c488","Type":"ContainerDied","Data":"a209638b76d4ea7161615ce8fafb3ab2b6292329d6043441f11c92a086f3285a"} Dec 03 16:15:45 crc kubenswrapper[4998]: I1203 16:15:45.556710 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wmnm9"] Dec 03 16:15:45 crc kubenswrapper[4998]: I1203 16:15:45.558356 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wmnm9" Dec 03 16:15:45 crc kubenswrapper[4998]: I1203 16:15:45.565951 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wmnm9"] Dec 03 16:15:45 crc kubenswrapper[4998]: I1203 16:15:45.632443 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8856963-919e-40d0-9014-4aea91403509-catalog-content\") pod \"redhat-operators-wmnm9\" (UID: \"a8856963-919e-40d0-9014-4aea91403509\") " pod="openshift-marketplace/redhat-operators-wmnm9" Dec 03 16:15:45 crc kubenswrapper[4998]: I1203 16:15:45.632529 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8856963-919e-40d0-9014-4aea91403509-utilities\") pod \"redhat-operators-wmnm9\" (UID: \"a8856963-919e-40d0-9014-4aea91403509\") " pod="openshift-marketplace/redhat-operators-wmnm9" Dec 03 16:15:45 crc kubenswrapper[4998]: I1203 16:15:45.632564 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6vf5\" (UniqueName: \"kubernetes.io/projected/a8856963-919e-40d0-9014-4aea91403509-kube-api-access-s6vf5\") pod \"redhat-operators-wmnm9\" (UID: \"a8856963-919e-40d0-9014-4aea91403509\") " pod="openshift-marketplace/redhat-operators-wmnm9" Dec 03 16:15:45 crc kubenswrapper[4998]: I1203 16:15:45.637689 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dtvp7" Dec 03 16:15:45 crc kubenswrapper[4998]: I1203 16:15:45.733869 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8856963-919e-40d0-9014-4aea91403509-utilities\") pod \"redhat-operators-wmnm9\" (UID: \"a8856963-919e-40d0-9014-4aea91403509\") " pod="openshift-marketplace/redhat-operators-wmnm9" Dec 03 16:15:45 crc kubenswrapper[4998]: I1203 16:15:45.733931 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6vf5\" (UniqueName: \"kubernetes.io/projected/a8856963-919e-40d0-9014-4aea91403509-kube-api-access-s6vf5\") pod \"redhat-operators-wmnm9\" (UID: \"a8856963-919e-40d0-9014-4aea91403509\") " pod="openshift-marketplace/redhat-operators-wmnm9" Dec 03 16:15:45 crc kubenswrapper[4998]: I1203 16:15:45.734057 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8856963-919e-40d0-9014-4aea91403509-catalog-content\") pod \"redhat-operators-wmnm9\" (UID: \"a8856963-919e-40d0-9014-4aea91403509\") " pod="openshift-marketplace/redhat-operators-wmnm9" Dec 03 16:15:45 crc kubenswrapper[4998]: I1203 16:15:45.734468 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8856963-919e-40d0-9014-4aea91403509-utilities\") pod \"redhat-operators-wmnm9\" (UID: \"a8856963-919e-40d0-9014-4aea91403509\") " pod="openshift-marketplace/redhat-operators-wmnm9" Dec 03 16:15:45 crc kubenswrapper[4998]: I1203 16:15:45.734582 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8856963-919e-40d0-9014-4aea91403509-catalog-content\") pod \"redhat-operators-wmnm9\" (UID: \"a8856963-919e-40d0-9014-4aea91403509\") " pod="openshift-marketplace/redhat-operators-wmnm9" Dec 03 16:15:45 crc kubenswrapper[4998]: I1203 16:15:45.753435 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6vf5\" (UniqueName: \"kubernetes.io/projected/a8856963-919e-40d0-9014-4aea91403509-kube-api-access-s6vf5\") pod \"redhat-operators-wmnm9\" (UID: \"a8856963-919e-40d0-9014-4aea91403509\") " pod="openshift-marketplace/redhat-operators-wmnm9" Dec 03 16:15:45 crc kubenswrapper[4998]: I1203 16:15:45.878015 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wmnm9" Dec 03 16:15:46 crc kubenswrapper[4998]: I1203 16:15:46.116976 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wmnm9"] Dec 03 16:15:46 crc kubenswrapper[4998]: W1203 16:15:46.130464 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8856963_919e_40d0_9014_4aea91403509.slice/crio-ea9c8282cf9ffd47e2d68e6a0f7808879a3e49e79e14f05c04a195c5979edf5a WatchSource:0}: Error finding container ea9c8282cf9ffd47e2d68e6a0f7808879a3e49e79e14f05c04a195c5979edf5a: Status 404 returned error can't find the container with id ea9c8282cf9ffd47e2d68e6a0f7808879a3e49e79e14f05c04a195c5979edf5a Dec 03 16:15:46 crc kubenswrapper[4998]: I1203 16:15:46.150337 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj" event={"ID":"9038bb91-6c7d-42c6-b864-9d522c73c488","Type":"ContainerStarted","Data":"0f5cc8f10e71db5ac5b8272af7536bcf99b22b49419e818272ffc932409761bd"} Dec 03 16:15:46 crc kubenswrapper[4998]: I1203 16:15:46.152551 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wmnm9" event={"ID":"a8856963-919e-40d0-9014-4aea91403509","Type":"ContainerStarted","Data":"ea9c8282cf9ffd47e2d68e6a0f7808879a3e49e79e14f05c04a195c5979edf5a"} Dec 03 16:15:47 crc kubenswrapper[4998]: I1203 16:15:47.161603 4998 generic.go:334] "Generic (PLEG): container finished" podID="a8856963-919e-40d0-9014-4aea91403509" containerID="ed4d7d1158a148300ff894f1c11e8f994cd94f2078c351d05c1b6f700b3ca921" exitCode=0 Dec 03 16:15:47 crc kubenswrapper[4998]: I1203 16:15:47.161702 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wmnm9" event={"ID":"a8856963-919e-40d0-9014-4aea91403509","Type":"ContainerDied","Data":"ed4d7d1158a148300ff894f1c11e8f994cd94f2078c351d05c1b6f700b3ca921"} Dec 03 16:15:47 crc kubenswrapper[4998]: I1203 16:15:47.164013 4998 generic.go:334] "Generic (PLEG): container finished" podID="9038bb91-6c7d-42c6-b864-9d522c73c488" containerID="0f5cc8f10e71db5ac5b8272af7536bcf99b22b49419e818272ffc932409761bd" exitCode=0 Dec 03 16:15:47 crc kubenswrapper[4998]: I1203 16:15:47.164074 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj" event={"ID":"9038bb91-6c7d-42c6-b864-9d522c73c488","Type":"ContainerDied","Data":"0f5cc8f10e71db5ac5b8272af7536bcf99b22b49419e818272ffc932409761bd"} Dec 03 16:15:48 crc kubenswrapper[4998]: I1203 16:15:48.172595 4998 generic.go:334] "Generic (PLEG): container finished" podID="9038bb91-6c7d-42c6-b864-9d522c73c488" containerID="96ecea7508d21d15322264ed535492daecb64b7812f05877ff79925c643d6c04" exitCode=0 Dec 03 16:15:48 crc kubenswrapper[4998]: I1203 16:15:48.172650 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj" event={"ID":"9038bb91-6c7d-42c6-b864-9d522c73c488","Type":"ContainerDied","Data":"96ecea7508d21d15322264ed535492daecb64b7812f05877ff79925c643d6c04"} Dec 03 16:15:48 crc kubenswrapper[4998]: I1203 16:15:48.176290 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wmnm9" event={"ID":"a8856963-919e-40d0-9014-4aea91403509","Type":"ContainerStarted","Data":"104d97deefbc8554b4c6b4fa37a45d66a9e91fb267955d99e7b9f51702354052"} Dec 03 16:15:49 crc kubenswrapper[4998]: I1203 16:15:49.186355 4998 generic.go:334] "Generic (PLEG): container finished" podID="a8856963-919e-40d0-9014-4aea91403509" containerID="104d97deefbc8554b4c6b4fa37a45d66a9e91fb267955d99e7b9f51702354052" exitCode=0 Dec 03 16:15:49 crc kubenswrapper[4998]: I1203 16:15:49.186449 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wmnm9" event={"ID":"a8856963-919e-40d0-9014-4aea91403509","Type":"ContainerDied","Data":"104d97deefbc8554b4c6b4fa37a45d66a9e91fb267955d99e7b9f51702354052"} Dec 03 16:15:49 crc kubenswrapper[4998]: I1203 16:15:49.490244 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj" Dec 03 16:15:49 crc kubenswrapper[4998]: I1203 16:15:49.580896 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9038bb91-6c7d-42c6-b864-9d522c73c488-util\") pod \"9038bb91-6c7d-42c6-b864-9d522c73c488\" (UID: \"9038bb91-6c7d-42c6-b864-9d522c73c488\") " Dec 03 16:15:49 crc kubenswrapper[4998]: I1203 16:15:49.580964 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdt2f\" (UniqueName: \"kubernetes.io/projected/9038bb91-6c7d-42c6-b864-9d522c73c488-kube-api-access-cdt2f\") pod \"9038bb91-6c7d-42c6-b864-9d522c73c488\" (UID: \"9038bb91-6c7d-42c6-b864-9d522c73c488\") " Dec 03 16:15:49 crc kubenswrapper[4998]: I1203 16:15:49.581048 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9038bb91-6c7d-42c6-b864-9d522c73c488-bundle\") pod \"9038bb91-6c7d-42c6-b864-9d522c73c488\" (UID: \"9038bb91-6c7d-42c6-b864-9d522c73c488\") " Dec 03 16:15:49 crc kubenswrapper[4998]: I1203 16:15:49.583276 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9038bb91-6c7d-42c6-b864-9d522c73c488-bundle" (OuterVolumeSpecName: "bundle") pod "9038bb91-6c7d-42c6-b864-9d522c73c488" (UID: "9038bb91-6c7d-42c6-b864-9d522c73c488"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:15:49 crc kubenswrapper[4998]: I1203 16:15:49.587135 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9038bb91-6c7d-42c6-b864-9d522c73c488-kube-api-access-cdt2f" (OuterVolumeSpecName: "kube-api-access-cdt2f") pod "9038bb91-6c7d-42c6-b864-9d522c73c488" (UID: "9038bb91-6c7d-42c6-b864-9d522c73c488"). InnerVolumeSpecName "kube-api-access-cdt2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:15:49 crc kubenswrapper[4998]: I1203 16:15:49.682725 4998 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9038bb91-6c7d-42c6-b864-9d522c73c488-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:49 crc kubenswrapper[4998]: I1203 16:15:49.682814 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdt2f\" (UniqueName: \"kubernetes.io/projected/9038bb91-6c7d-42c6-b864-9d522c73c488-kube-api-access-cdt2f\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:49 crc kubenswrapper[4998]: I1203 16:15:49.906179 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9038bb91-6c7d-42c6-b864-9d522c73c488-util" (OuterVolumeSpecName: "util") pod "9038bb91-6c7d-42c6-b864-9d522c73c488" (UID: "9038bb91-6c7d-42c6-b864-9d522c73c488"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:15:49 crc kubenswrapper[4998]: I1203 16:15:49.988062 4998 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9038bb91-6c7d-42c6-b864-9d522c73c488-util\") on node \"crc\" DevicePath \"\"" Dec 03 16:15:50 crc kubenswrapper[4998]: I1203 16:15:50.200465 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj" event={"ID":"9038bb91-6c7d-42c6-b864-9d522c73c488","Type":"ContainerDied","Data":"bcd420968c3b8bf20ecb3c9c328563e9149f89baee492ef2ad9cf30fb1666209"} Dec 03 16:15:50 crc kubenswrapper[4998]: I1203 16:15:50.200516 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bcd420968c3b8bf20ecb3c9c328563e9149f89baee492ef2ad9cf30fb1666209" Dec 03 16:15:50 crc kubenswrapper[4998]: I1203 16:15:50.200514 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj" Dec 03 16:15:50 crc kubenswrapper[4998]: I1203 16:15:50.203863 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wmnm9" event={"ID":"a8856963-919e-40d0-9014-4aea91403509","Type":"ContainerStarted","Data":"6dcea555602bf7ee666c489c36027d62b9e5c0d01b49b01aa727d72efb541e5f"} Dec 03 16:15:50 crc kubenswrapper[4998]: I1203 16:15:50.236829 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wmnm9" podStartSLOduration=2.44155379 podStartE2EDuration="5.236807378s" podCreationTimestamp="2025-12-03 16:15:45 +0000 UTC" firstStartedPulling="2025-12-03 16:15:47.163904925 +0000 UTC m=+725.775605158" lastFinishedPulling="2025-12-03 16:15:49.959158493 +0000 UTC m=+728.570858746" observedRunningTime="2025-12-03 16:15:50.225879857 +0000 UTC m=+728.837580150" watchObservedRunningTime="2025-12-03 16:15:50.236807378 +0000 UTC m=+728.848507621" Dec 03 16:15:55 crc kubenswrapper[4998]: I1203 16:15:55.879336 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wmnm9" Dec 03 16:15:55 crc kubenswrapper[4998]: I1203 16:15:55.880477 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wmnm9" Dec 03 16:15:56 crc kubenswrapper[4998]: I1203 16:15:56.925025 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wmnm9" podUID="a8856963-919e-40d0-9014-4aea91403509" containerName="registry-server" probeResult="failure" output=< Dec 03 16:15:56 crc kubenswrapper[4998]: timeout: failed to connect service ":50051" within 1s Dec 03 16:15:56 crc kubenswrapper[4998]: > Dec 03 16:15:57 crc kubenswrapper[4998]: I1203 16:15:57.111284 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:15:57 crc kubenswrapper[4998]: I1203 16:15:57.111345 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.279481 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-f8jt6"] Dec 03 16:15:58 crc kubenswrapper[4998]: E1203 16:15:58.279712 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9038bb91-6c7d-42c6-b864-9d522c73c488" containerName="util" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.279726 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="9038bb91-6c7d-42c6-b864-9d522c73c488" containerName="util" Dec 03 16:15:58 crc kubenswrapper[4998]: E1203 16:15:58.279738 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9038bb91-6c7d-42c6-b864-9d522c73c488" containerName="pull" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.279745 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="9038bb91-6c7d-42c6-b864-9d522c73c488" containerName="pull" Dec 03 16:15:58 crc kubenswrapper[4998]: E1203 16:15:58.279796 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9038bb91-6c7d-42c6-b864-9d522c73c488" containerName="extract" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.279811 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="9038bb91-6c7d-42c6-b864-9d522c73c488" containerName="extract" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.279956 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="9038bb91-6c7d-42c6-b864-9d522c73c488" containerName="extract" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.280360 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-f8jt6" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.282278 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.282685 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-f69lj" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.282734 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.300009 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-f8jt6"] Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.328656 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-4ffgs"] Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.329509 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-4ffgs" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.331685 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-f6jbf" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.334281 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.342826 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-d46s4"] Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.343543 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-d46s4" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.351303 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-4ffgs"] Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.373433 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-d46s4"] Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.390282 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6a29f86a-9f1e-44b6-a1e2-1cdc7dd3e7ec-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5c4dbf9785-4ffgs\" (UID: \"6a29f86a-9f1e-44b6-a1e2-1cdc7dd3e7ec\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-4ffgs" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.390336 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0dedded0-9a7b-4697-9a69-6ca56016f0f1-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5c4dbf9785-d46s4\" (UID: \"0dedded0-9a7b-4697-9a69-6ca56016f0f1\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-d46s4" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.390364 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6a29f86a-9f1e-44b6-a1e2-1cdc7dd3e7ec-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5c4dbf9785-4ffgs\" (UID: \"6a29f86a-9f1e-44b6-a1e2-1cdc7dd3e7ec\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-4ffgs" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.390383 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9qdc\" (UniqueName: \"kubernetes.io/projected/50b3e173-2cde-4732-9fa4-0d5cc7936dcd-kube-api-access-l9qdc\") pod \"obo-prometheus-operator-668cf9dfbb-f8jt6\" (UID: \"50b3e173-2cde-4732-9fa4-0d5cc7936dcd\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-f8jt6" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.390464 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0dedded0-9a7b-4697-9a69-6ca56016f0f1-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5c4dbf9785-d46s4\" (UID: \"0dedded0-9a7b-4697-9a69-6ca56016f0f1\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-d46s4" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.491772 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6a29f86a-9f1e-44b6-a1e2-1cdc7dd3e7ec-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5c4dbf9785-4ffgs\" (UID: \"6a29f86a-9f1e-44b6-a1e2-1cdc7dd3e7ec\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-4ffgs" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.491858 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0dedded0-9a7b-4697-9a69-6ca56016f0f1-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5c4dbf9785-d46s4\" (UID: \"0dedded0-9a7b-4697-9a69-6ca56016f0f1\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-d46s4" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.491896 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6a29f86a-9f1e-44b6-a1e2-1cdc7dd3e7ec-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5c4dbf9785-4ffgs\" (UID: \"6a29f86a-9f1e-44b6-a1e2-1cdc7dd3e7ec\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-4ffgs" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.491929 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9qdc\" (UniqueName: \"kubernetes.io/projected/50b3e173-2cde-4732-9fa4-0d5cc7936dcd-kube-api-access-l9qdc\") pod \"obo-prometheus-operator-668cf9dfbb-f8jt6\" (UID: \"50b3e173-2cde-4732-9fa4-0d5cc7936dcd\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-f8jt6" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.491960 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0dedded0-9a7b-4697-9a69-6ca56016f0f1-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5c4dbf9785-d46s4\" (UID: \"0dedded0-9a7b-4697-9a69-6ca56016f0f1\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-d46s4" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.499039 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6a29f86a-9f1e-44b6-a1e2-1cdc7dd3e7ec-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5c4dbf9785-4ffgs\" (UID: \"6a29f86a-9f1e-44b6-a1e2-1cdc7dd3e7ec\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-4ffgs" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.499709 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0dedded0-9a7b-4697-9a69-6ca56016f0f1-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5c4dbf9785-d46s4\" (UID: \"0dedded0-9a7b-4697-9a69-6ca56016f0f1\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-d46s4" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.500038 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6a29f86a-9f1e-44b6-a1e2-1cdc7dd3e7ec-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5c4dbf9785-4ffgs\" (UID: \"6a29f86a-9f1e-44b6-a1e2-1cdc7dd3e7ec\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-4ffgs" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.501205 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0dedded0-9a7b-4697-9a69-6ca56016f0f1-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5c4dbf9785-d46s4\" (UID: \"0dedded0-9a7b-4697-9a69-6ca56016f0f1\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-d46s4" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.510859 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9qdc\" (UniqueName: \"kubernetes.io/projected/50b3e173-2cde-4732-9fa4-0d5cc7936dcd-kube-api-access-l9qdc\") pod \"obo-prometheus-operator-668cf9dfbb-f8jt6\" (UID: \"50b3e173-2cde-4732-9fa4-0d5cc7936dcd\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-f8jt6" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.526919 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-zzvhx"] Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.527710 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-zzvhx" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.529543 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-nbvpd" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.530850 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.541258 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-zzvhx"] Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.593474 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/60293ca8-98bb-49a8-b54f-0c3aa3f84cc3-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-zzvhx\" (UID: \"60293ca8-98bb-49a8-b54f-0c3aa3f84cc3\") " pod="openshift-operators/observability-operator-d8bb48f5d-zzvhx" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.593553 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-775zb\" (UniqueName: \"kubernetes.io/projected/60293ca8-98bb-49a8-b54f-0c3aa3f84cc3-kube-api-access-775zb\") pod \"observability-operator-d8bb48f5d-zzvhx\" (UID: \"60293ca8-98bb-49a8-b54f-0c3aa3f84cc3\") " pod="openshift-operators/observability-operator-d8bb48f5d-zzvhx" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.595339 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-f8jt6" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.623300 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-2wk7t"] Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.623970 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-2wk7t" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.626774 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-xm9vb" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.638040 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-2wk7t"] Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.648835 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-4ffgs" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.660930 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-d46s4" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.694274 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/60293ca8-98bb-49a8-b54f-0c3aa3f84cc3-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-zzvhx\" (UID: \"60293ca8-98bb-49a8-b54f-0c3aa3f84cc3\") " pod="openshift-operators/observability-operator-d8bb48f5d-zzvhx" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.694356 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/48acd5a8-2b63-4683-9f39-43f3e3193a85-openshift-service-ca\") pod \"perses-operator-5446b9c989-2wk7t\" (UID: \"48acd5a8-2b63-4683-9f39-43f3e3193a85\") " pod="openshift-operators/perses-operator-5446b9c989-2wk7t" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.694379 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-775zb\" (UniqueName: \"kubernetes.io/projected/60293ca8-98bb-49a8-b54f-0c3aa3f84cc3-kube-api-access-775zb\") pod \"observability-operator-d8bb48f5d-zzvhx\" (UID: \"60293ca8-98bb-49a8-b54f-0c3aa3f84cc3\") " pod="openshift-operators/observability-operator-d8bb48f5d-zzvhx" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.694406 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qv6bn\" (UniqueName: \"kubernetes.io/projected/48acd5a8-2b63-4683-9f39-43f3e3193a85-kube-api-access-qv6bn\") pod \"perses-operator-5446b9c989-2wk7t\" (UID: \"48acd5a8-2b63-4683-9f39-43f3e3193a85\") " pod="openshift-operators/perses-operator-5446b9c989-2wk7t" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.698103 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/60293ca8-98bb-49a8-b54f-0c3aa3f84cc3-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-zzvhx\" (UID: \"60293ca8-98bb-49a8-b54f-0c3aa3f84cc3\") " pod="openshift-operators/observability-operator-d8bb48f5d-zzvhx" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.718885 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-775zb\" (UniqueName: \"kubernetes.io/projected/60293ca8-98bb-49a8-b54f-0c3aa3f84cc3-kube-api-access-775zb\") pod \"observability-operator-d8bb48f5d-zzvhx\" (UID: \"60293ca8-98bb-49a8-b54f-0c3aa3f84cc3\") " pod="openshift-operators/observability-operator-d8bb48f5d-zzvhx" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.796306 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/48acd5a8-2b63-4683-9f39-43f3e3193a85-openshift-service-ca\") pod \"perses-operator-5446b9c989-2wk7t\" (UID: \"48acd5a8-2b63-4683-9f39-43f3e3193a85\") " pod="openshift-operators/perses-operator-5446b9c989-2wk7t" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.796583 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qv6bn\" (UniqueName: \"kubernetes.io/projected/48acd5a8-2b63-4683-9f39-43f3e3193a85-kube-api-access-qv6bn\") pod \"perses-operator-5446b9c989-2wk7t\" (UID: \"48acd5a8-2b63-4683-9f39-43f3e3193a85\") " pod="openshift-operators/perses-operator-5446b9c989-2wk7t" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.798394 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/48acd5a8-2b63-4683-9f39-43f3e3193a85-openshift-service-ca\") pod \"perses-operator-5446b9c989-2wk7t\" (UID: \"48acd5a8-2b63-4683-9f39-43f3e3193a85\") " pod="openshift-operators/perses-operator-5446b9c989-2wk7t" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.824426 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qv6bn\" (UniqueName: \"kubernetes.io/projected/48acd5a8-2b63-4683-9f39-43f3e3193a85-kube-api-access-qv6bn\") pod \"perses-operator-5446b9c989-2wk7t\" (UID: \"48acd5a8-2b63-4683-9f39-43f3e3193a85\") " pod="openshift-operators/perses-operator-5446b9c989-2wk7t" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.852098 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-f8jt6"] Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.866011 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-zzvhx" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.951157 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-4ffgs"] Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.959001 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-2wk7t" Dec 03 16:15:58 crc kubenswrapper[4998]: I1203 16:15:58.991555 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-d46s4"] Dec 03 16:15:59 crc kubenswrapper[4998]: W1203 16:15:59.000904 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0dedded0_9a7b_4697_9a69_6ca56016f0f1.slice/crio-9db7981fd559a5a1a93a5c3968cd63d61016742f653a2355353d642a4c056f46 WatchSource:0}: Error finding container 9db7981fd559a5a1a93a5c3968cd63d61016742f653a2355353d642a4c056f46: Status 404 returned error can't find the container with id 9db7981fd559a5a1a93a5c3968cd63d61016742f653a2355353d642a4c056f46 Dec 03 16:15:59 crc kubenswrapper[4998]: I1203 16:15:59.071349 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-zzvhx"] Dec 03 16:15:59 crc kubenswrapper[4998]: W1203 16:15:59.083291 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60293ca8_98bb_49a8_b54f_0c3aa3f84cc3.slice/crio-5b33ad4c39291d1a07bf43403b8a7b2a1b693944df7b7f3dfb83479eb97b28ed WatchSource:0}: Error finding container 5b33ad4c39291d1a07bf43403b8a7b2a1b693944df7b7f3dfb83479eb97b28ed: Status 404 returned error can't find the container with id 5b33ad4c39291d1a07bf43403b8a7b2a1b693944df7b7f3dfb83479eb97b28ed Dec 03 16:15:59 crc kubenswrapper[4998]: I1203 16:15:59.177316 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-2wk7t"] Dec 03 16:15:59 crc kubenswrapper[4998]: I1203 16:15:59.625513 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-f8jt6" event={"ID":"50b3e173-2cde-4732-9fa4-0d5cc7936dcd","Type":"ContainerStarted","Data":"e87f75b4d789876667fe921900840043eb171976f4eff3197affae913778a584"} Dec 03 16:15:59 crc kubenswrapper[4998]: I1203 16:15:59.626930 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-4ffgs" event={"ID":"6a29f86a-9f1e-44b6-a1e2-1cdc7dd3e7ec","Type":"ContainerStarted","Data":"ee63f0bbefbb977aa3df5adc91e86ae5956bc2d2cc50ad0251270c81a9541675"} Dec 03 16:15:59 crc kubenswrapper[4998]: I1203 16:15:59.628039 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-d46s4" event={"ID":"0dedded0-9a7b-4697-9a69-6ca56016f0f1","Type":"ContainerStarted","Data":"9db7981fd559a5a1a93a5c3968cd63d61016742f653a2355353d642a4c056f46"} Dec 03 16:15:59 crc kubenswrapper[4998]: I1203 16:15:59.629443 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-zzvhx" event={"ID":"60293ca8-98bb-49a8-b54f-0c3aa3f84cc3","Type":"ContainerStarted","Data":"5b33ad4c39291d1a07bf43403b8a7b2a1b693944df7b7f3dfb83479eb97b28ed"} Dec 03 16:15:59 crc kubenswrapper[4998]: I1203 16:15:59.630784 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-2wk7t" event={"ID":"48acd5a8-2b63-4683-9f39-43f3e3193a85","Type":"ContainerStarted","Data":"ed146c679f7fea30a34680c2c170d3c1a1d2c305f4733598e5a361b126b86a73"} Dec 03 16:16:05 crc kubenswrapper[4998]: I1203 16:16:05.929409 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wmnm9" Dec 03 16:16:05 crc kubenswrapper[4998]: I1203 16:16:05.991626 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wmnm9" Dec 03 16:16:06 crc kubenswrapper[4998]: I1203 16:16:06.164026 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wmnm9"] Dec 03 16:16:07 crc kubenswrapper[4998]: I1203 16:16:07.674653 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wmnm9" podUID="a8856963-919e-40d0-9014-4aea91403509" containerName="registry-server" containerID="cri-o://6dcea555602bf7ee666c489c36027d62b9e5c0d01b49b01aa727d72efb541e5f" gracePeriod=2 Dec 03 16:16:08 crc kubenswrapper[4998]: I1203 16:16:08.681572 4998 generic.go:334] "Generic (PLEG): container finished" podID="a8856963-919e-40d0-9014-4aea91403509" containerID="6dcea555602bf7ee666c489c36027d62b9e5c0d01b49b01aa727d72efb541e5f" exitCode=0 Dec 03 16:16:08 crc kubenswrapper[4998]: I1203 16:16:08.681618 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wmnm9" event={"ID":"a8856963-919e-40d0-9014-4aea91403509","Type":"ContainerDied","Data":"6dcea555602bf7ee666c489c36027d62b9e5c0d01b49b01aa727d72efb541e5f"} Dec 03 16:16:15 crc kubenswrapper[4998]: E1203 16:16:15.026876 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec" Dec 03 16:16:15 crc kubenswrapper[4998]: E1203 16:16:15.027360 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator-admission-webhook,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec,Command:[],Args:[--web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{209715200 0} {} BinarySI},},Requests:ResourceList{cpu: {{50 -3} {} 50m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-admission-webhook-5c4dbf9785-d46s4_openshift-operators(0dedded0-9a7b-4697-9a69-6ca56016f0f1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 16:16:15 crc kubenswrapper[4998]: E1203 16:16:15.028775 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-d46s4" podUID="0dedded0-9a7b-4697-9a69-6ca56016f0f1" Dec 03 16:16:15 crc kubenswrapper[4998]: E1203 16:16:15.067647 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec" Dec 03 16:16:15 crc kubenswrapper[4998]: E1203 16:16:15.067817 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator-admission-webhook,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec,Command:[],Args:[--web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{209715200 0} {} BinarySI},},Requests:ResourceList{cpu: {{50 -3} {} 50m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-admission-webhook-5c4dbf9785-4ffgs_openshift-operators(6a29f86a-9f1e-44b6-a1e2-1cdc7dd3e7ec): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 16:16:15 crc kubenswrapper[4998]: E1203 16:16:15.069007 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-4ffgs" podUID="6a29f86a-9f1e-44b6-a1e2-1cdc7dd3e7ec" Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.083834 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wmnm9" Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.109163 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8856963-919e-40d0-9014-4aea91403509-catalog-content\") pod \"a8856963-919e-40d0-9014-4aea91403509\" (UID: \"a8856963-919e-40d0-9014-4aea91403509\") " Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.109239 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6vf5\" (UniqueName: \"kubernetes.io/projected/a8856963-919e-40d0-9014-4aea91403509-kube-api-access-s6vf5\") pod \"a8856963-919e-40d0-9014-4aea91403509\" (UID: \"a8856963-919e-40d0-9014-4aea91403509\") " Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.109314 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8856963-919e-40d0-9014-4aea91403509-utilities\") pod \"a8856963-919e-40d0-9014-4aea91403509\" (UID: \"a8856963-919e-40d0-9014-4aea91403509\") " Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.110542 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8856963-919e-40d0-9014-4aea91403509-utilities" (OuterVolumeSpecName: "utilities") pod "a8856963-919e-40d0-9014-4aea91403509" (UID: "a8856963-919e-40d0-9014-4aea91403509"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.127655 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8856963-919e-40d0-9014-4aea91403509-kube-api-access-s6vf5" (OuterVolumeSpecName: "kube-api-access-s6vf5") pod "a8856963-919e-40d0-9014-4aea91403509" (UID: "a8856963-919e-40d0-9014-4aea91403509"). InnerVolumeSpecName "kube-api-access-s6vf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.210467 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8856963-919e-40d0-9014-4aea91403509-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.210502 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6vf5\" (UniqueName: \"kubernetes.io/projected/a8856963-919e-40d0-9014-4aea91403509-kube-api-access-s6vf5\") on node \"crc\" DevicePath \"\"" Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.230487 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8856963-919e-40d0-9014-4aea91403509-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a8856963-919e-40d0-9014-4aea91403509" (UID: "a8856963-919e-40d0-9014-4aea91403509"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.311437 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8856963-919e-40d0-9014-4aea91403509-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.722736 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wmnm9" event={"ID":"a8856963-919e-40d0-9014-4aea91403509","Type":"ContainerDied","Data":"ea9c8282cf9ffd47e2d68e6a0f7808879a3e49e79e14f05c04a195c5979edf5a"} Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.723003 4998 scope.go:117] "RemoveContainer" containerID="6dcea555602bf7ee666c489c36027d62b9e5c0d01b49b01aa727d72efb541e5f" Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.722850 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wmnm9" Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.724693 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-zzvhx" event={"ID":"60293ca8-98bb-49a8-b54f-0c3aa3f84cc3","Type":"ContainerStarted","Data":"5b38abaa8b8d2d1d494b5690a515ff6b1f26c6bc00503b514e973ddf5670f7be"} Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.725170 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-zzvhx" Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.727151 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-2wk7t" event={"ID":"48acd5a8-2b63-4683-9f39-43f3e3193a85","Type":"ContainerStarted","Data":"3202774b249365ef2df5e07185d86fa5af80f1197a7fbe34eac986b39ee836ef"} Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.727263 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-2wk7t" Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.729938 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-f8jt6" event={"ID":"50b3e173-2cde-4732-9fa4-0d5cc7936dcd","Type":"ContainerStarted","Data":"a477513ba04e63f269d1b42aa09c48208f0ae86194f5297320aaec9bcbb8bc09"} Dec 03 16:16:15 crc kubenswrapper[4998]: E1203 16:16:15.730579 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec\\\"\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-4ffgs" podUID="6a29f86a-9f1e-44b6-a1e2-1cdc7dd3e7ec" Dec 03 16:16:15 crc kubenswrapper[4998]: E1203 16:16:15.730749 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec\\\"\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-d46s4" podUID="0dedded0-9a7b-4697-9a69-6ca56016f0f1" Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.742990 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wmnm9"] Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.745577 4998 scope.go:117] "RemoveContainer" containerID="104d97deefbc8554b4c6b4fa37a45d66a9e91fb267955d99e7b9f51702354052" Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.746679 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wmnm9"] Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.760808 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-zzvhx" podStartSLOduration=1.756522063 podStartE2EDuration="17.760792541s" podCreationTimestamp="2025-12-03 16:15:58 +0000 UTC" firstStartedPulling="2025-12-03 16:15:59.086330747 +0000 UTC m=+737.698030970" lastFinishedPulling="2025-12-03 16:16:15.090601215 +0000 UTC m=+753.702301448" observedRunningTime="2025-12-03 16:16:15.75803716 +0000 UTC m=+754.369737393" watchObservedRunningTime="2025-12-03 16:16:15.760792541 +0000 UTC m=+754.372492754" Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.762419 4998 scope.go:117] "RemoveContainer" containerID="ed4d7d1158a148300ff894f1c11e8f994cd94f2078c351d05c1b6f700b3ca921" Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.786854 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-zzvhx" Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.796883 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-2wk7t" podStartSLOduration=1.9125908489999999 podStartE2EDuration="17.796862679s" podCreationTimestamp="2025-12-03 16:15:58 +0000 UTC" firstStartedPulling="2025-12-03 16:15:59.20381648 +0000 UTC m=+737.815516703" lastFinishedPulling="2025-12-03 16:16:15.08808831 +0000 UTC m=+753.699788533" observedRunningTime="2025-12-03 16:16:15.77902143 +0000 UTC m=+754.390721653" watchObservedRunningTime="2025-12-03 16:16:15.796862679 +0000 UTC m=+754.408562902" Dec 03 16:16:15 crc kubenswrapper[4998]: I1203 16:16:15.815849 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-f8jt6" podStartSLOduration=1.628850567 podStartE2EDuration="17.815831017s" podCreationTimestamp="2025-12-03 16:15:58 +0000 UTC" firstStartedPulling="2025-12-03 16:15:58.880683025 +0000 UTC m=+737.492383258" lastFinishedPulling="2025-12-03 16:16:15.067663465 +0000 UTC m=+753.679363708" observedRunningTime="2025-12-03 16:16:15.812803569 +0000 UTC m=+754.424503792" watchObservedRunningTime="2025-12-03 16:16:15.815831017 +0000 UTC m=+754.427531240" Dec 03 16:16:17 crc kubenswrapper[4998]: I1203 16:16:17.688686 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8856963-919e-40d0-9014-4aea91403509" path="/var/lib/kubelet/pods/a8856963-919e-40d0-9014-4aea91403509/volumes" Dec 03 16:16:27 crc kubenswrapper[4998]: I1203 16:16:27.111909 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:16:27 crc kubenswrapper[4998]: I1203 16:16:27.112566 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:16:27 crc kubenswrapper[4998]: I1203 16:16:27.112631 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:16:27 crc kubenswrapper[4998]: I1203 16:16:27.113490 4998 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bc9f09c91a152bc17df6a34b6b589619602b656e78ba0e9011a364c51abd030e"} pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 16:16:27 crc kubenswrapper[4998]: I1203 16:16:27.113586 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" containerID="cri-o://bc9f09c91a152bc17df6a34b6b589619602b656e78ba0e9011a364c51abd030e" gracePeriod=600 Dec 03 16:16:27 crc kubenswrapper[4998]: I1203 16:16:27.801193 4998 generic.go:334] "Generic (PLEG): container finished" podID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerID="bc9f09c91a152bc17df6a34b6b589619602b656e78ba0e9011a364c51abd030e" exitCode=0 Dec 03 16:16:27 crc kubenswrapper[4998]: I1203 16:16:27.801277 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerDied","Data":"bc9f09c91a152bc17df6a34b6b589619602b656e78ba0e9011a364c51abd030e"} Dec 03 16:16:27 crc kubenswrapper[4998]: I1203 16:16:27.801476 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"5b97e0b62564fae4effbf36e0feb1d3ce5f2c4a4d5ac7c086146fd76fc8ee224"} Dec 03 16:16:27 crc kubenswrapper[4998]: I1203 16:16:27.801502 4998 scope.go:117] "RemoveContainer" containerID="947673c25a57749d81b7b602dfa5d8e08e3e29d4682222a096f3e504d993caf4" Dec 03 16:16:28 crc kubenswrapper[4998]: I1203 16:16:28.961247 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-2wk7t" Dec 03 16:16:29 crc kubenswrapper[4998]: I1203 16:16:29.817221 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-d46s4" event={"ID":"0dedded0-9a7b-4697-9a69-6ca56016f0f1","Type":"ContainerStarted","Data":"340e9322bbb1b30246b4c8324aba3f07903e0e0a3642780d28a13c30979db880"} Dec 03 16:16:29 crc kubenswrapper[4998]: I1203 16:16:29.840145 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-d46s4" podStartSLOduration=1.356621171 podStartE2EDuration="31.840124525s" podCreationTimestamp="2025-12-03 16:15:58 +0000 UTC" firstStartedPulling="2025-12-03 16:15:59.005746663 +0000 UTC m=+737.617446886" lastFinishedPulling="2025-12-03 16:16:29.489250007 +0000 UTC m=+768.100950240" observedRunningTime="2025-12-03 16:16:29.836694397 +0000 UTC m=+768.448394610" watchObservedRunningTime="2025-12-03 16:16:29.840124525 +0000 UTC m=+768.451824768" Dec 03 16:16:30 crc kubenswrapper[4998]: I1203 16:16:30.825070 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-4ffgs" event={"ID":"6a29f86a-9f1e-44b6-a1e2-1cdc7dd3e7ec","Type":"ContainerStarted","Data":"37e3a35addbf3cf07fd2886f62bd7ae53a2c59f8595dc470ad04dfd39bf3d42d"} Dec 03 16:16:30 crc kubenswrapper[4998]: I1203 16:16:30.853184 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c4dbf9785-4ffgs" podStartSLOduration=-9223372004.00161 podStartE2EDuration="32.853166764s" podCreationTimestamp="2025-12-03 16:15:58 +0000 UTC" firstStartedPulling="2025-12-03 16:15:58.97338048 +0000 UTC m=+737.585080703" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:16:30.85070352 +0000 UTC m=+769.462403753" watchObservedRunningTime="2025-12-03 16:16:30.853166764 +0000 UTC m=+769.464866997" Dec 03 16:16:48 crc kubenswrapper[4998]: I1203 16:16:48.157807 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l"] Dec 03 16:16:48 crc kubenswrapper[4998]: E1203 16:16:48.158730 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8856963-919e-40d0-9014-4aea91403509" containerName="registry-server" Dec 03 16:16:48 crc kubenswrapper[4998]: I1203 16:16:48.158751 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8856963-919e-40d0-9014-4aea91403509" containerName="registry-server" Dec 03 16:16:48 crc kubenswrapper[4998]: E1203 16:16:48.158800 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8856963-919e-40d0-9014-4aea91403509" containerName="extract-utilities" Dec 03 16:16:48 crc kubenswrapper[4998]: I1203 16:16:48.158813 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8856963-919e-40d0-9014-4aea91403509" containerName="extract-utilities" Dec 03 16:16:48 crc kubenswrapper[4998]: E1203 16:16:48.158833 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8856963-919e-40d0-9014-4aea91403509" containerName="extract-content" Dec 03 16:16:48 crc kubenswrapper[4998]: I1203 16:16:48.158846 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8856963-919e-40d0-9014-4aea91403509" containerName="extract-content" Dec 03 16:16:48 crc kubenswrapper[4998]: I1203 16:16:48.159028 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8856963-919e-40d0-9014-4aea91403509" containerName="registry-server" Dec 03 16:16:48 crc kubenswrapper[4998]: I1203 16:16:48.160260 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l" Dec 03 16:16:48 crc kubenswrapper[4998]: I1203 16:16:48.162840 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 16:16:48 crc kubenswrapper[4998]: I1203 16:16:48.174309 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l"] Dec 03 16:16:48 crc kubenswrapper[4998]: I1203 16:16:48.310325 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc94f\" (UniqueName: \"kubernetes.io/projected/b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99-kube-api-access-rc94f\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l\" (UID: \"b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l" Dec 03 16:16:48 crc kubenswrapper[4998]: I1203 16:16:48.310446 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l\" (UID: \"b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l" Dec 03 16:16:48 crc kubenswrapper[4998]: I1203 16:16:48.310475 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l\" (UID: \"b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l" Dec 03 16:16:48 crc kubenswrapper[4998]: I1203 16:16:48.411281 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l\" (UID: \"b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l" Dec 03 16:16:48 crc kubenswrapper[4998]: I1203 16:16:48.411385 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc94f\" (UniqueName: \"kubernetes.io/projected/b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99-kube-api-access-rc94f\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l\" (UID: \"b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l" Dec 03 16:16:48 crc kubenswrapper[4998]: I1203 16:16:48.411463 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l\" (UID: \"b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l" Dec 03 16:16:48 crc kubenswrapper[4998]: I1203 16:16:48.412030 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l\" (UID: \"b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l" Dec 03 16:16:48 crc kubenswrapper[4998]: I1203 16:16:48.412113 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l\" (UID: \"b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l" Dec 03 16:16:48 crc kubenswrapper[4998]: I1203 16:16:48.429277 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc94f\" (UniqueName: \"kubernetes.io/projected/b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99-kube-api-access-rc94f\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l\" (UID: \"b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l" Dec 03 16:16:48 crc kubenswrapper[4998]: I1203 16:16:48.482170 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l" Dec 03 16:16:48 crc kubenswrapper[4998]: I1203 16:16:48.900729 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l"] Dec 03 16:16:48 crc kubenswrapper[4998]: W1203 16:16:48.916940 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7bba5b4_55d7_465f_89a3_3ed3b1cd0d99.slice/crio-5befc1e43bec6508ae95f6cfbd6aff28200a4a05f6f1d2aef6f3605a6532f311 WatchSource:0}: Error finding container 5befc1e43bec6508ae95f6cfbd6aff28200a4a05f6f1d2aef6f3605a6532f311: Status 404 returned error can't find the container with id 5befc1e43bec6508ae95f6cfbd6aff28200a4a05f6f1d2aef6f3605a6532f311 Dec 03 16:16:48 crc kubenswrapper[4998]: I1203 16:16:48.951336 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l" event={"ID":"b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99","Type":"ContainerStarted","Data":"5befc1e43bec6508ae95f6cfbd6aff28200a4a05f6f1d2aef6f3605a6532f311"} Dec 03 16:16:51 crc kubenswrapper[4998]: I1203 16:16:51.970855 4998 generic.go:334] "Generic (PLEG): container finished" podID="b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99" containerID="bcd7c2b654b7377ec6e9d07abe9c2ef5882d2a52d2a6da235cdede24e221b4f8" exitCode=0 Dec 03 16:16:51 crc kubenswrapper[4998]: I1203 16:16:51.971081 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l" event={"ID":"b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99","Type":"ContainerDied","Data":"bcd7c2b654b7377ec6e9d07abe9c2ef5882d2a52d2a6da235cdede24e221b4f8"} Dec 03 16:16:54 crc kubenswrapper[4998]: I1203 16:16:54.995326 4998 generic.go:334] "Generic (PLEG): container finished" podID="b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99" containerID="4fea203177d7e2dfdd5f0fd41e861bf7e87b47b3850ab2f78a26ec5e702735cc" exitCode=0 Dec 03 16:16:54 crc kubenswrapper[4998]: I1203 16:16:54.995386 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l" event={"ID":"b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99","Type":"ContainerDied","Data":"4fea203177d7e2dfdd5f0fd41e861bf7e87b47b3850ab2f78a26ec5e702735cc"} Dec 03 16:16:56 crc kubenswrapper[4998]: I1203 16:16:56.006827 4998 generic.go:334] "Generic (PLEG): container finished" podID="b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99" containerID="e7800075cb485aca4393152bfbdbd6baee5f1212ab51ab5c6754ac27f5d91061" exitCode=0 Dec 03 16:16:56 crc kubenswrapper[4998]: I1203 16:16:56.006925 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l" event={"ID":"b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99","Type":"ContainerDied","Data":"e7800075cb485aca4393152bfbdbd6baee5f1212ab51ab5c6754ac27f5d91061"} Dec 03 16:16:57 crc kubenswrapper[4998]: I1203 16:16:57.297146 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l" Dec 03 16:16:57 crc kubenswrapper[4998]: I1203 16:16:57.429750 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99-bundle\") pod \"b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99\" (UID: \"b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99\") " Dec 03 16:16:57 crc kubenswrapper[4998]: I1203 16:16:57.429855 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rc94f\" (UniqueName: \"kubernetes.io/projected/b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99-kube-api-access-rc94f\") pod \"b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99\" (UID: \"b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99\") " Dec 03 16:16:57 crc kubenswrapper[4998]: I1203 16:16:57.429926 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99-util\") pod \"b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99\" (UID: \"b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99\") " Dec 03 16:16:57 crc kubenswrapper[4998]: I1203 16:16:57.430683 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99-bundle" (OuterVolumeSpecName: "bundle") pod "b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99" (UID: "b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:16:57 crc kubenswrapper[4998]: I1203 16:16:57.435428 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99-kube-api-access-rc94f" (OuterVolumeSpecName: "kube-api-access-rc94f") pod "b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99" (UID: "b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99"). InnerVolumeSpecName "kube-api-access-rc94f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:16:57 crc kubenswrapper[4998]: I1203 16:16:57.451121 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99-util" (OuterVolumeSpecName: "util") pod "b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99" (UID: "b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:16:57 crc kubenswrapper[4998]: I1203 16:16:57.531555 4998 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:16:57 crc kubenswrapper[4998]: I1203 16:16:57.531595 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rc94f\" (UniqueName: \"kubernetes.io/projected/b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99-kube-api-access-rc94f\") on node \"crc\" DevicePath \"\"" Dec 03 16:16:57 crc kubenswrapper[4998]: I1203 16:16:57.531609 4998 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99-util\") on node \"crc\" DevicePath \"\"" Dec 03 16:16:58 crc kubenswrapper[4998]: I1203 16:16:58.023026 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l" event={"ID":"b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99","Type":"ContainerDied","Data":"5befc1e43bec6508ae95f6cfbd6aff28200a4a05f6f1d2aef6f3605a6532f311"} Dec 03 16:16:58 crc kubenswrapper[4998]: I1203 16:16:58.023072 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5befc1e43bec6508ae95f6cfbd6aff28200a4a05f6f1d2aef6f3605a6532f311" Dec 03 16:16:58 crc kubenswrapper[4998]: I1203 16:16:58.023085 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l" Dec 03 16:17:00 crc kubenswrapper[4998]: I1203 16:17:00.308886 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-cdghz"] Dec 03 16:17:00 crc kubenswrapper[4998]: E1203 16:17:00.309700 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99" containerName="util" Dec 03 16:17:00 crc kubenswrapper[4998]: I1203 16:17:00.309728 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99" containerName="util" Dec 03 16:17:00 crc kubenswrapper[4998]: E1203 16:17:00.309791 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99" containerName="pull" Dec 03 16:17:00 crc kubenswrapper[4998]: I1203 16:17:00.309813 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99" containerName="pull" Dec 03 16:17:00 crc kubenswrapper[4998]: E1203 16:17:00.309851 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99" containerName="extract" Dec 03 16:17:00 crc kubenswrapper[4998]: I1203 16:17:00.309868 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99" containerName="extract" Dec 03 16:17:00 crc kubenswrapper[4998]: I1203 16:17:00.310111 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99" containerName="extract" Dec 03 16:17:00 crc kubenswrapper[4998]: I1203 16:17:00.311009 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-cdghz" Dec 03 16:17:00 crc kubenswrapper[4998]: I1203 16:17:00.313215 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 03 16:17:00 crc kubenswrapper[4998]: I1203 16:17:00.314720 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 03 16:17:00 crc kubenswrapper[4998]: I1203 16:17:00.315285 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-sdl5j" Dec 03 16:17:00 crc kubenswrapper[4998]: I1203 16:17:00.317140 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-cdghz"] Dec 03 16:17:00 crc kubenswrapper[4998]: I1203 16:17:00.468388 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbtm7\" (UniqueName: \"kubernetes.io/projected/2f9b5504-1b4e-4ca7-8fc7-1319edc59ce9-kube-api-access-rbtm7\") pod \"nmstate-operator-5b5b58f5c8-cdghz\" (UID: \"2f9b5504-1b4e-4ca7-8fc7-1319edc59ce9\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-cdghz" Dec 03 16:17:00 crc kubenswrapper[4998]: I1203 16:17:00.569795 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbtm7\" (UniqueName: \"kubernetes.io/projected/2f9b5504-1b4e-4ca7-8fc7-1319edc59ce9-kube-api-access-rbtm7\") pod \"nmstate-operator-5b5b58f5c8-cdghz\" (UID: \"2f9b5504-1b4e-4ca7-8fc7-1319edc59ce9\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-cdghz" Dec 03 16:17:00 crc kubenswrapper[4998]: I1203 16:17:00.605159 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbtm7\" (UniqueName: \"kubernetes.io/projected/2f9b5504-1b4e-4ca7-8fc7-1319edc59ce9-kube-api-access-rbtm7\") pod \"nmstate-operator-5b5b58f5c8-cdghz\" (UID: \"2f9b5504-1b4e-4ca7-8fc7-1319edc59ce9\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-cdghz" Dec 03 16:17:00 crc kubenswrapper[4998]: I1203 16:17:00.628394 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-cdghz" Dec 03 16:17:00 crc kubenswrapper[4998]: I1203 16:17:00.847092 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-cdghz"] Dec 03 16:17:01 crc kubenswrapper[4998]: I1203 16:17:01.039780 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-cdghz" event={"ID":"2f9b5504-1b4e-4ca7-8fc7-1319edc59ce9","Type":"ContainerStarted","Data":"5da7ead2ace432bb420022b0d3fb0934924771f4e3582fe49341041816fcb879"} Dec 03 16:17:05 crc kubenswrapper[4998]: I1203 16:17:05.067744 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-cdghz" event={"ID":"2f9b5504-1b4e-4ca7-8fc7-1319edc59ce9","Type":"ContainerStarted","Data":"54a364f24daf943d7d1b57e3b1a3c789d06e2ffbc11361c6b491bbbf5decc311"} Dec 03 16:17:05 crc kubenswrapper[4998]: I1203 16:17:05.103031 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-cdghz" podStartSLOduration=1.989978757 podStartE2EDuration="5.103001012s" podCreationTimestamp="2025-12-03 16:17:00 +0000 UTC" firstStartedPulling="2025-12-03 16:17:00.854843847 +0000 UTC m=+799.466544070" lastFinishedPulling="2025-12-03 16:17:03.967866102 +0000 UTC m=+802.579566325" observedRunningTime="2025-12-03 16:17:05.091595099 +0000 UTC m=+803.703295372" watchObservedRunningTime="2025-12-03 16:17:05.103001012 +0000 UTC m=+803.714701265" Dec 03 16:17:08 crc kubenswrapper[4998]: I1203 16:17:08.093367 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5bc7h"] Dec 03 16:17:08 crc kubenswrapper[4998]: I1203 16:17:08.096484 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5bc7h" Dec 03 16:17:08 crc kubenswrapper[4998]: I1203 16:17:08.101022 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5bc7h"] Dec 03 16:17:08 crc kubenswrapper[4998]: I1203 16:17:08.259571 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32577841-0e29-442c-b7e3-492335ab2d4c-utilities\") pod \"redhat-marketplace-5bc7h\" (UID: \"32577841-0e29-442c-b7e3-492335ab2d4c\") " pod="openshift-marketplace/redhat-marketplace-5bc7h" Dec 03 16:17:08 crc kubenswrapper[4998]: I1203 16:17:08.259668 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32577841-0e29-442c-b7e3-492335ab2d4c-catalog-content\") pod \"redhat-marketplace-5bc7h\" (UID: \"32577841-0e29-442c-b7e3-492335ab2d4c\") " pod="openshift-marketplace/redhat-marketplace-5bc7h" Dec 03 16:17:08 crc kubenswrapper[4998]: I1203 16:17:08.259709 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psw76\" (UniqueName: \"kubernetes.io/projected/32577841-0e29-442c-b7e3-492335ab2d4c-kube-api-access-psw76\") pod \"redhat-marketplace-5bc7h\" (UID: \"32577841-0e29-442c-b7e3-492335ab2d4c\") " pod="openshift-marketplace/redhat-marketplace-5bc7h" Dec 03 16:17:08 crc kubenswrapper[4998]: I1203 16:17:08.360744 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32577841-0e29-442c-b7e3-492335ab2d4c-utilities\") pod \"redhat-marketplace-5bc7h\" (UID: \"32577841-0e29-442c-b7e3-492335ab2d4c\") " pod="openshift-marketplace/redhat-marketplace-5bc7h" Dec 03 16:17:08 crc kubenswrapper[4998]: I1203 16:17:08.360936 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32577841-0e29-442c-b7e3-492335ab2d4c-catalog-content\") pod \"redhat-marketplace-5bc7h\" (UID: \"32577841-0e29-442c-b7e3-492335ab2d4c\") " pod="openshift-marketplace/redhat-marketplace-5bc7h" Dec 03 16:17:08 crc kubenswrapper[4998]: I1203 16:17:08.360982 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psw76\" (UniqueName: \"kubernetes.io/projected/32577841-0e29-442c-b7e3-492335ab2d4c-kube-api-access-psw76\") pod \"redhat-marketplace-5bc7h\" (UID: \"32577841-0e29-442c-b7e3-492335ab2d4c\") " pod="openshift-marketplace/redhat-marketplace-5bc7h" Dec 03 16:17:08 crc kubenswrapper[4998]: I1203 16:17:08.361514 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32577841-0e29-442c-b7e3-492335ab2d4c-utilities\") pod \"redhat-marketplace-5bc7h\" (UID: \"32577841-0e29-442c-b7e3-492335ab2d4c\") " pod="openshift-marketplace/redhat-marketplace-5bc7h" Dec 03 16:17:08 crc kubenswrapper[4998]: I1203 16:17:08.361943 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32577841-0e29-442c-b7e3-492335ab2d4c-catalog-content\") pod \"redhat-marketplace-5bc7h\" (UID: \"32577841-0e29-442c-b7e3-492335ab2d4c\") " pod="openshift-marketplace/redhat-marketplace-5bc7h" Dec 03 16:17:08 crc kubenswrapper[4998]: I1203 16:17:08.394860 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psw76\" (UniqueName: \"kubernetes.io/projected/32577841-0e29-442c-b7e3-492335ab2d4c-kube-api-access-psw76\") pod \"redhat-marketplace-5bc7h\" (UID: \"32577841-0e29-442c-b7e3-492335ab2d4c\") " pod="openshift-marketplace/redhat-marketplace-5bc7h" Dec 03 16:17:08 crc kubenswrapper[4998]: I1203 16:17:08.417345 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5bc7h" Dec 03 16:17:08 crc kubenswrapper[4998]: I1203 16:17:08.854877 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5bc7h"] Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.057656 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-n9fkd"] Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.061446 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-n9fkd" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.065971 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-hsqqv" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.076936 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-n9fkd"] Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.090603 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7rcw5"] Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.091322 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7rcw5" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.107146 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.122855 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7rcw5"] Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.123156 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5bc7h" event={"ID":"32577841-0e29-442c-b7e3-492335ab2d4c","Type":"ContainerStarted","Data":"a37e82548404bde528367b0111941471fe8ade2d8eaa3c74f2564a9c00e83364"} Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.123253 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5bc7h" event={"ID":"32577841-0e29-442c-b7e3-492335ab2d4c","Type":"ContainerStarted","Data":"48b2677e89acb69569386b27c01efbda514b786d9a9474574d44e7ddd28184b3"} Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.127527 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-plkwj"] Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.128597 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-plkwj" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.171647 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5mkk\" (UniqueName: \"kubernetes.io/projected/c74a341f-037c-472d-af3c-d76406b0f422-kube-api-access-g5mkk\") pod \"nmstate-metrics-7f946cbc9-n9fkd\" (UID: \"c74a341f-037c-472d-af3c-d76406b0f422\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-n9fkd" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.211199 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5td8z"] Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.211854 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5td8z" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.221381 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.221564 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-66slg" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.221678 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.224132 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5td8z"] Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.273192 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/4a388be2-7ba3-4fdd-896f-28577a929261-nmstate-lock\") pod \"nmstate-handler-plkwj\" (UID: \"4a388be2-7ba3-4fdd-896f-28577a929261\") " pod="openshift-nmstate/nmstate-handler-plkwj" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.273270 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/4a388be2-7ba3-4fdd-896f-28577a929261-ovs-socket\") pod \"nmstate-handler-plkwj\" (UID: \"4a388be2-7ba3-4fdd-896f-28577a929261\") " pod="openshift-nmstate/nmstate-handler-plkwj" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.273370 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccl8q\" (UniqueName: \"kubernetes.io/projected/bf4a690b-a2ab-41b9-a6d6-f053d2c4d735-kube-api-access-ccl8q\") pod \"nmstate-webhook-5f6d4c5ccb-7rcw5\" (UID: \"bf4a690b-a2ab-41b9-a6d6-f053d2c4d735\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7rcw5" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.273426 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4ckw\" (UniqueName: \"kubernetes.io/projected/4a388be2-7ba3-4fdd-896f-28577a929261-kube-api-access-m4ckw\") pod \"nmstate-handler-plkwj\" (UID: \"4a388be2-7ba3-4fdd-896f-28577a929261\") " pod="openshift-nmstate/nmstate-handler-plkwj" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.273525 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bf4a690b-a2ab-41b9-a6d6-f053d2c4d735-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-7rcw5\" (UID: \"bf4a690b-a2ab-41b9-a6d6-f053d2c4d735\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7rcw5" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.273547 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/4a388be2-7ba3-4fdd-896f-28577a929261-dbus-socket\") pod \"nmstate-handler-plkwj\" (UID: \"4a388be2-7ba3-4fdd-896f-28577a929261\") " pod="openshift-nmstate/nmstate-handler-plkwj" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.273599 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5mkk\" (UniqueName: \"kubernetes.io/projected/c74a341f-037c-472d-af3c-d76406b0f422-kube-api-access-g5mkk\") pod \"nmstate-metrics-7f946cbc9-n9fkd\" (UID: \"c74a341f-037c-472d-af3c-d76406b0f422\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-n9fkd" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.299057 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5mkk\" (UniqueName: \"kubernetes.io/projected/c74a341f-037c-472d-af3c-d76406b0f422-kube-api-access-g5mkk\") pod \"nmstate-metrics-7f946cbc9-n9fkd\" (UID: \"c74a341f-037c-472d-af3c-d76406b0f422\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-n9fkd" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.374692 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4ckw\" (UniqueName: \"kubernetes.io/projected/4a388be2-7ba3-4fdd-896f-28577a929261-kube-api-access-m4ckw\") pod \"nmstate-handler-plkwj\" (UID: \"4a388be2-7ba3-4fdd-896f-28577a929261\") " pod="openshift-nmstate/nmstate-handler-plkwj" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.374778 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99hc7\" (UniqueName: \"kubernetes.io/projected/9d6e2533-1816-4ea6-a1cc-dc7f834e1c33-kube-api-access-99hc7\") pod \"nmstate-console-plugin-7fbb5f6569-5td8z\" (UID: \"9d6e2533-1816-4ea6-a1cc-dc7f834e1c33\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5td8z" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.374803 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bf4a690b-a2ab-41b9-a6d6-f053d2c4d735-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-7rcw5\" (UID: \"bf4a690b-a2ab-41b9-a6d6-f053d2c4d735\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7rcw5" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.374817 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/4a388be2-7ba3-4fdd-896f-28577a929261-dbus-socket\") pod \"nmstate-handler-plkwj\" (UID: \"4a388be2-7ba3-4fdd-896f-28577a929261\") " pod="openshift-nmstate/nmstate-handler-plkwj" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.374847 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/4a388be2-7ba3-4fdd-896f-28577a929261-nmstate-lock\") pod \"nmstate-handler-plkwj\" (UID: \"4a388be2-7ba3-4fdd-896f-28577a929261\") " pod="openshift-nmstate/nmstate-handler-plkwj" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.374865 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/4a388be2-7ba3-4fdd-896f-28577a929261-ovs-socket\") pod \"nmstate-handler-plkwj\" (UID: \"4a388be2-7ba3-4fdd-896f-28577a929261\") " pod="openshift-nmstate/nmstate-handler-plkwj" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.374885 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/9d6e2533-1816-4ea6-a1cc-dc7f834e1c33-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-5td8z\" (UID: \"9d6e2533-1816-4ea6-a1cc-dc7f834e1c33\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5td8z" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.374907 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d6e2533-1816-4ea6-a1cc-dc7f834e1c33-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-5td8z\" (UID: \"9d6e2533-1816-4ea6-a1cc-dc7f834e1c33\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5td8z" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.374925 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccl8q\" (UniqueName: \"kubernetes.io/projected/bf4a690b-a2ab-41b9-a6d6-f053d2c4d735-kube-api-access-ccl8q\") pod \"nmstate-webhook-5f6d4c5ccb-7rcw5\" (UID: \"bf4a690b-a2ab-41b9-a6d6-f053d2c4d735\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7rcw5" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.375995 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/4a388be2-7ba3-4fdd-896f-28577a929261-ovs-socket\") pod \"nmstate-handler-plkwj\" (UID: \"4a388be2-7ba3-4fdd-896f-28577a929261\") " pod="openshift-nmstate/nmstate-handler-plkwj" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.376047 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/4a388be2-7ba3-4fdd-896f-28577a929261-nmstate-lock\") pod \"nmstate-handler-plkwj\" (UID: \"4a388be2-7ba3-4fdd-896f-28577a929261\") " pod="openshift-nmstate/nmstate-handler-plkwj" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.376065 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/4a388be2-7ba3-4fdd-896f-28577a929261-dbus-socket\") pod \"nmstate-handler-plkwj\" (UID: \"4a388be2-7ba3-4fdd-896f-28577a929261\") " pod="openshift-nmstate/nmstate-handler-plkwj" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.380733 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bf4a690b-a2ab-41b9-a6d6-f053d2c4d735-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-7rcw5\" (UID: \"bf4a690b-a2ab-41b9-a6d6-f053d2c4d735\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7rcw5" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.385001 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-n9fkd" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.392372 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccl8q\" (UniqueName: \"kubernetes.io/projected/bf4a690b-a2ab-41b9-a6d6-f053d2c4d735-kube-api-access-ccl8q\") pod \"nmstate-webhook-5f6d4c5ccb-7rcw5\" (UID: \"bf4a690b-a2ab-41b9-a6d6-f053d2c4d735\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7rcw5" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.394198 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4ckw\" (UniqueName: \"kubernetes.io/projected/4a388be2-7ba3-4fdd-896f-28577a929261-kube-api-access-m4ckw\") pod \"nmstate-handler-plkwj\" (UID: \"4a388be2-7ba3-4fdd-896f-28577a929261\") " pod="openshift-nmstate/nmstate-handler-plkwj" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.413867 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-789646f4f5-xq522"] Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.414622 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.426142 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-789646f4f5-xq522"] Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.455956 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7rcw5" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.465388 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-plkwj" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.493449 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8da0e557-cc5d-4fe2-8df5-7f13d70d13c6-console-serving-cert\") pod \"console-789646f4f5-xq522\" (UID: \"8da0e557-cc5d-4fe2-8df5-7f13d70d13c6\") " pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.493476 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8da0e557-cc5d-4fe2-8df5-7f13d70d13c6-console-config\") pod \"console-789646f4f5-xq522\" (UID: \"8da0e557-cc5d-4fe2-8df5-7f13d70d13c6\") " pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.493501 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8da0e557-cc5d-4fe2-8df5-7f13d70d13c6-service-ca\") pod \"console-789646f4f5-xq522\" (UID: \"8da0e557-cc5d-4fe2-8df5-7f13d70d13c6\") " pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.493532 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99hc7\" (UniqueName: \"kubernetes.io/projected/9d6e2533-1816-4ea6-a1cc-dc7f834e1c33-kube-api-access-99hc7\") pod \"nmstate-console-plugin-7fbb5f6569-5td8z\" (UID: \"9d6e2533-1816-4ea6-a1cc-dc7f834e1c33\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5td8z" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.493574 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8da0e557-cc5d-4fe2-8df5-7f13d70d13c6-oauth-serving-cert\") pod \"console-789646f4f5-xq522\" (UID: \"8da0e557-cc5d-4fe2-8df5-7f13d70d13c6\") " pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.493592 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z7z9\" (UniqueName: \"kubernetes.io/projected/8da0e557-cc5d-4fe2-8df5-7f13d70d13c6-kube-api-access-8z7z9\") pod \"console-789646f4f5-xq522\" (UID: \"8da0e557-cc5d-4fe2-8df5-7f13d70d13c6\") " pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.493612 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/9d6e2533-1816-4ea6-a1cc-dc7f834e1c33-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-5td8z\" (UID: \"9d6e2533-1816-4ea6-a1cc-dc7f834e1c33\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5td8z" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.493635 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d6e2533-1816-4ea6-a1cc-dc7f834e1c33-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-5td8z\" (UID: \"9d6e2533-1816-4ea6-a1cc-dc7f834e1c33\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5td8z" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.493659 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8da0e557-cc5d-4fe2-8df5-7f13d70d13c6-trusted-ca-bundle\") pod \"console-789646f4f5-xq522\" (UID: \"8da0e557-cc5d-4fe2-8df5-7f13d70d13c6\") " pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.493677 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8da0e557-cc5d-4fe2-8df5-7f13d70d13c6-console-oauth-config\") pod \"console-789646f4f5-xq522\" (UID: \"8da0e557-cc5d-4fe2-8df5-7f13d70d13c6\") " pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.495327 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/9d6e2533-1816-4ea6-a1cc-dc7f834e1c33-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-5td8z\" (UID: \"9d6e2533-1816-4ea6-a1cc-dc7f834e1c33\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5td8z" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.500027 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d6e2533-1816-4ea6-a1cc-dc7f834e1c33-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-5td8z\" (UID: \"9d6e2533-1816-4ea6-a1cc-dc7f834e1c33\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5td8z" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.508316 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99hc7\" (UniqueName: \"kubernetes.io/projected/9d6e2533-1816-4ea6-a1cc-dc7f834e1c33-kube-api-access-99hc7\") pod \"nmstate-console-plugin-7fbb5f6569-5td8z\" (UID: \"9d6e2533-1816-4ea6-a1cc-dc7f834e1c33\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5td8z" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.534192 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5td8z" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.594436 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8da0e557-cc5d-4fe2-8df5-7f13d70d13c6-oauth-serving-cert\") pod \"console-789646f4f5-xq522\" (UID: \"8da0e557-cc5d-4fe2-8df5-7f13d70d13c6\") " pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.594478 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z7z9\" (UniqueName: \"kubernetes.io/projected/8da0e557-cc5d-4fe2-8df5-7f13d70d13c6-kube-api-access-8z7z9\") pod \"console-789646f4f5-xq522\" (UID: \"8da0e557-cc5d-4fe2-8df5-7f13d70d13c6\") " pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.594515 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8da0e557-cc5d-4fe2-8df5-7f13d70d13c6-trusted-ca-bundle\") pod \"console-789646f4f5-xq522\" (UID: \"8da0e557-cc5d-4fe2-8df5-7f13d70d13c6\") " pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.594541 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8da0e557-cc5d-4fe2-8df5-7f13d70d13c6-console-oauth-config\") pod \"console-789646f4f5-xq522\" (UID: \"8da0e557-cc5d-4fe2-8df5-7f13d70d13c6\") " pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.594559 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8da0e557-cc5d-4fe2-8df5-7f13d70d13c6-console-serving-cert\") pod \"console-789646f4f5-xq522\" (UID: \"8da0e557-cc5d-4fe2-8df5-7f13d70d13c6\") " pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.594575 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8da0e557-cc5d-4fe2-8df5-7f13d70d13c6-console-config\") pod \"console-789646f4f5-xq522\" (UID: \"8da0e557-cc5d-4fe2-8df5-7f13d70d13c6\") " pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.594598 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8da0e557-cc5d-4fe2-8df5-7f13d70d13c6-service-ca\") pod \"console-789646f4f5-xq522\" (UID: \"8da0e557-cc5d-4fe2-8df5-7f13d70d13c6\") " pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.595579 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8da0e557-cc5d-4fe2-8df5-7f13d70d13c6-service-ca\") pod \"console-789646f4f5-xq522\" (UID: \"8da0e557-cc5d-4fe2-8df5-7f13d70d13c6\") " pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.595693 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8da0e557-cc5d-4fe2-8df5-7f13d70d13c6-oauth-serving-cert\") pod \"console-789646f4f5-xq522\" (UID: \"8da0e557-cc5d-4fe2-8df5-7f13d70d13c6\") " pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.596472 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8da0e557-cc5d-4fe2-8df5-7f13d70d13c6-console-config\") pod \"console-789646f4f5-xq522\" (UID: \"8da0e557-cc5d-4fe2-8df5-7f13d70d13c6\") " pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.597075 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8da0e557-cc5d-4fe2-8df5-7f13d70d13c6-trusted-ca-bundle\") pod \"console-789646f4f5-xq522\" (UID: \"8da0e557-cc5d-4fe2-8df5-7f13d70d13c6\") " pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.600821 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8da0e557-cc5d-4fe2-8df5-7f13d70d13c6-console-oauth-config\") pod \"console-789646f4f5-xq522\" (UID: \"8da0e557-cc5d-4fe2-8df5-7f13d70d13c6\") " pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.601453 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8da0e557-cc5d-4fe2-8df5-7f13d70d13c6-console-serving-cert\") pod \"console-789646f4f5-xq522\" (UID: \"8da0e557-cc5d-4fe2-8df5-7f13d70d13c6\") " pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.613967 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z7z9\" (UniqueName: \"kubernetes.io/projected/8da0e557-cc5d-4fe2-8df5-7f13d70d13c6-kube-api-access-8z7z9\") pod \"console-789646f4f5-xq522\" (UID: \"8da0e557-cc5d-4fe2-8df5-7f13d70d13c6\") " pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.688641 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7rcw5"] Dec 03 16:17:09 crc kubenswrapper[4998]: W1203 16:17:09.689599 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf4a690b_a2ab_41b9_a6d6_f053d2c4d735.slice/crio-3eca4335552561fae3574a18f2db92fda70e3c77890afdc47b1c5387b235ba5c WatchSource:0}: Error finding container 3eca4335552561fae3574a18f2db92fda70e3c77890afdc47b1c5387b235ba5c: Status 404 returned error can't find the container with id 3eca4335552561fae3574a18f2db92fda70e3c77890afdc47b1c5387b235ba5c Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.733892 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5td8z"] Dec 03 16:17:09 crc kubenswrapper[4998]: W1203 16:17:09.736193 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d6e2533_1816_4ea6_a1cc_dc7f834e1c33.slice/crio-1ff314962b7f6b85e83b6384c6eb5e5582689232d6a908caaa2322327c610f9d WatchSource:0}: Error finding container 1ff314962b7f6b85e83b6384c6eb5e5582689232d6a908caaa2322327c610f9d: Status 404 returned error can't find the container with id 1ff314962b7f6b85e83b6384c6eb5e5582689232d6a908caaa2322327c610f9d Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.764343 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.816127 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-n9fkd"] Dec 03 16:17:09 crc kubenswrapper[4998]: W1203 16:17:09.817423 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc74a341f_037c_472d_af3c_d76406b0f422.slice/crio-ba4eeab95efd9124c1c0118d11fe906c2f012f860c8666057290e668084904ba WatchSource:0}: Error finding container ba4eeab95efd9124c1c0118d11fe906c2f012f860c8666057290e668084904ba: Status 404 returned error can't find the container with id ba4eeab95efd9124c1c0118d11fe906c2f012f860c8666057290e668084904ba Dec 03 16:17:09 crc kubenswrapper[4998]: I1203 16:17:09.934068 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-789646f4f5-xq522"] Dec 03 16:17:09 crc kubenswrapper[4998]: W1203 16:17:09.939370 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8da0e557_cc5d_4fe2_8df5_7f13d70d13c6.slice/crio-e0a816cce22c089ff90d777c709405a010cf368ac094a8eb013abf09b348a2bd WatchSource:0}: Error finding container e0a816cce22c089ff90d777c709405a010cf368ac094a8eb013abf09b348a2bd: Status 404 returned error can't find the container with id e0a816cce22c089ff90d777c709405a010cf368ac094a8eb013abf09b348a2bd Dec 03 16:17:10 crc kubenswrapper[4998]: I1203 16:17:10.131959 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-n9fkd" event={"ID":"c74a341f-037c-472d-af3c-d76406b0f422","Type":"ContainerStarted","Data":"ba4eeab95efd9124c1c0118d11fe906c2f012f860c8666057290e668084904ba"} Dec 03 16:17:10 crc kubenswrapper[4998]: I1203 16:17:10.133573 4998 generic.go:334] "Generic (PLEG): container finished" podID="32577841-0e29-442c-b7e3-492335ab2d4c" containerID="a37e82548404bde528367b0111941471fe8ade2d8eaa3c74f2564a9c00e83364" exitCode=0 Dec 03 16:17:10 crc kubenswrapper[4998]: I1203 16:17:10.133709 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5bc7h" event={"ID":"32577841-0e29-442c-b7e3-492335ab2d4c","Type":"ContainerDied","Data":"a37e82548404bde528367b0111941471fe8ade2d8eaa3c74f2564a9c00e83364"} Dec 03 16:17:10 crc kubenswrapper[4998]: I1203 16:17:10.135645 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-plkwj" event={"ID":"4a388be2-7ba3-4fdd-896f-28577a929261","Type":"ContainerStarted","Data":"3ad985210ef12e72eed868a09679494145547667ecd90038acf1a1c02ca95086"} Dec 03 16:17:10 crc kubenswrapper[4998]: I1203 16:17:10.137007 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5td8z" event={"ID":"9d6e2533-1816-4ea6-a1cc-dc7f834e1c33","Type":"ContainerStarted","Data":"1ff314962b7f6b85e83b6384c6eb5e5582689232d6a908caaa2322327c610f9d"} Dec 03 16:17:10 crc kubenswrapper[4998]: I1203 16:17:10.138699 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7rcw5" event={"ID":"bf4a690b-a2ab-41b9-a6d6-f053d2c4d735","Type":"ContainerStarted","Data":"3eca4335552561fae3574a18f2db92fda70e3c77890afdc47b1c5387b235ba5c"} Dec 03 16:17:10 crc kubenswrapper[4998]: I1203 16:17:10.140514 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-789646f4f5-xq522" event={"ID":"8da0e557-cc5d-4fe2-8df5-7f13d70d13c6","Type":"ContainerStarted","Data":"e0a816cce22c089ff90d777c709405a010cf368ac094a8eb013abf09b348a2bd"} Dec 03 16:17:11 crc kubenswrapper[4998]: I1203 16:17:11.154276 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-789646f4f5-xq522" event={"ID":"8da0e557-cc5d-4fe2-8df5-7f13d70d13c6","Type":"ContainerStarted","Data":"266aa75859808e832dc39e3fbd9640762ead4da4698e633386b403e6df262e37"} Dec 03 16:17:11 crc kubenswrapper[4998]: I1203 16:17:11.156374 4998 generic.go:334] "Generic (PLEG): container finished" podID="32577841-0e29-442c-b7e3-492335ab2d4c" containerID="5aafef86b40f875aac0d5b7eff13fbae6abd320dda74a9be51beab331a9f936f" exitCode=0 Dec 03 16:17:11 crc kubenswrapper[4998]: I1203 16:17:11.156403 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5bc7h" event={"ID":"32577841-0e29-442c-b7e3-492335ab2d4c","Type":"ContainerDied","Data":"5aafef86b40f875aac0d5b7eff13fbae6abd320dda74a9be51beab331a9f936f"} Dec 03 16:17:11 crc kubenswrapper[4998]: I1203 16:17:11.173655 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-789646f4f5-xq522" podStartSLOduration=2.173638544 podStartE2EDuration="2.173638544s" podCreationTimestamp="2025-12-03 16:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:17:11.170411381 +0000 UTC m=+809.782111604" watchObservedRunningTime="2025-12-03 16:17:11.173638544 +0000 UTC m=+809.785338757" Dec 03 16:17:13 crc kubenswrapper[4998]: I1203 16:17:13.173972 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5td8z" event={"ID":"9d6e2533-1816-4ea6-a1cc-dc7f834e1c33","Type":"ContainerStarted","Data":"f0f6e157e0364903fed89eb4457a8054230b0e622c805bffbcdabbe33a5ef40e"} Dec 03 16:17:13 crc kubenswrapper[4998]: I1203 16:17:13.176822 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7rcw5" event={"ID":"bf4a690b-a2ab-41b9-a6d6-f053d2c4d735","Type":"ContainerStarted","Data":"9e01f2e6d25a398e456fe6ce754c2b46fac91cfb7c50fcea43d92b319bcfdac3"} Dec 03 16:17:13 crc kubenswrapper[4998]: I1203 16:17:13.177073 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7rcw5" Dec 03 16:17:13 crc kubenswrapper[4998]: I1203 16:17:13.179302 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-n9fkd" event={"ID":"c74a341f-037c-472d-af3c-d76406b0f422","Type":"ContainerStarted","Data":"6b7eb7701587d0beb573f41dc35a04ad32ab19934393e6140fb4c4007237be7c"} Dec 03 16:17:13 crc kubenswrapper[4998]: I1203 16:17:13.181999 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5bc7h" event={"ID":"32577841-0e29-442c-b7e3-492335ab2d4c","Type":"ContainerStarted","Data":"b3ebef1adeb796121bf8c68d788da8039e5276ff66a6a27a610b90eb8183e691"} Dec 03 16:17:13 crc kubenswrapper[4998]: I1203 16:17:13.183808 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-plkwj" event={"ID":"4a388be2-7ba3-4fdd-896f-28577a929261","Type":"ContainerStarted","Data":"7b5285eb9a28e61e951408c158c98f971c59a52b6e7c37f1b82c9f3a9e28b114"} Dec 03 16:17:13 crc kubenswrapper[4998]: I1203 16:17:13.184111 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-plkwj" Dec 03 16:17:13 crc kubenswrapper[4998]: I1203 16:17:13.195195 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5td8z" podStartSLOduration=1.309576901 podStartE2EDuration="4.195173604s" podCreationTimestamp="2025-12-03 16:17:09 +0000 UTC" firstStartedPulling="2025-12-03 16:17:09.738072514 +0000 UTC m=+808.349772737" lastFinishedPulling="2025-12-03 16:17:12.623669197 +0000 UTC m=+811.235369440" observedRunningTime="2025-12-03 16:17:13.18724662 +0000 UTC m=+811.798946853" watchObservedRunningTime="2025-12-03 16:17:13.195173604 +0000 UTC m=+811.806873827" Dec 03 16:17:13 crc kubenswrapper[4998]: I1203 16:17:13.288143 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-plkwj" podStartSLOduration=1.13857439 podStartE2EDuration="4.288120905s" podCreationTimestamp="2025-12-03 16:17:09 +0000 UTC" firstStartedPulling="2025-12-03 16:17:09.488274926 +0000 UTC m=+808.099975149" lastFinishedPulling="2025-12-03 16:17:12.637821401 +0000 UTC m=+811.249521664" observedRunningTime="2025-12-03 16:17:13.236151808 +0000 UTC m=+811.847852021" watchObservedRunningTime="2025-12-03 16:17:13.288120905 +0000 UTC m=+811.899821128" Dec 03 16:17:13 crc kubenswrapper[4998]: I1203 16:17:13.311450 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7rcw5" podStartSLOduration=1.391921419 podStartE2EDuration="4.311427745s" podCreationTimestamp="2025-12-03 16:17:09 +0000 UTC" firstStartedPulling="2025-12-03 16:17:09.696886904 +0000 UTC m=+808.308587127" lastFinishedPulling="2025-12-03 16:17:12.61639319 +0000 UTC m=+811.228093453" observedRunningTime="2025-12-03 16:17:13.27626631 +0000 UTC m=+811.887966533" watchObservedRunningTime="2025-12-03 16:17:13.311427745 +0000 UTC m=+811.923127968" Dec 03 16:17:13 crc kubenswrapper[4998]: I1203 16:17:13.329384 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5bc7h" podStartSLOduration=2.858044355 podStartE2EDuration="5.329363287s" podCreationTimestamp="2025-12-03 16:17:08 +0000 UTC" firstStartedPulling="2025-12-03 16:17:10.135565523 +0000 UTC m=+808.747265746" lastFinishedPulling="2025-12-03 16:17:12.606884455 +0000 UTC m=+811.218584678" observedRunningTime="2025-12-03 16:17:13.312027211 +0000 UTC m=+811.923727464" watchObservedRunningTime="2025-12-03 16:17:13.329363287 +0000 UTC m=+811.941063510" Dec 03 16:17:14 crc kubenswrapper[4998]: I1203 16:17:14.301154 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mr2xt"] Dec 03 16:17:14 crc kubenswrapper[4998]: I1203 16:17:14.303004 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mr2xt" Dec 03 16:17:14 crc kubenswrapper[4998]: I1203 16:17:14.307704 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mr2xt"] Dec 03 16:17:14 crc kubenswrapper[4998]: I1203 16:17:14.459909 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e966d3c5-b228-4842-a3d6-9d9c9de192ce-catalog-content\") pod \"certified-operators-mr2xt\" (UID: \"e966d3c5-b228-4842-a3d6-9d9c9de192ce\") " pod="openshift-marketplace/certified-operators-mr2xt" Dec 03 16:17:14 crc kubenswrapper[4998]: I1203 16:17:14.459967 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbzdn\" (UniqueName: \"kubernetes.io/projected/e966d3c5-b228-4842-a3d6-9d9c9de192ce-kube-api-access-zbzdn\") pod \"certified-operators-mr2xt\" (UID: \"e966d3c5-b228-4842-a3d6-9d9c9de192ce\") " pod="openshift-marketplace/certified-operators-mr2xt" Dec 03 16:17:14 crc kubenswrapper[4998]: I1203 16:17:14.459987 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e966d3c5-b228-4842-a3d6-9d9c9de192ce-utilities\") pod \"certified-operators-mr2xt\" (UID: \"e966d3c5-b228-4842-a3d6-9d9c9de192ce\") " pod="openshift-marketplace/certified-operators-mr2xt" Dec 03 16:17:14 crc kubenswrapper[4998]: I1203 16:17:14.561281 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e966d3c5-b228-4842-a3d6-9d9c9de192ce-catalog-content\") pod \"certified-operators-mr2xt\" (UID: \"e966d3c5-b228-4842-a3d6-9d9c9de192ce\") " pod="openshift-marketplace/certified-operators-mr2xt" Dec 03 16:17:14 crc kubenswrapper[4998]: I1203 16:17:14.561347 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbzdn\" (UniqueName: \"kubernetes.io/projected/e966d3c5-b228-4842-a3d6-9d9c9de192ce-kube-api-access-zbzdn\") pod \"certified-operators-mr2xt\" (UID: \"e966d3c5-b228-4842-a3d6-9d9c9de192ce\") " pod="openshift-marketplace/certified-operators-mr2xt" Dec 03 16:17:14 crc kubenswrapper[4998]: I1203 16:17:14.561367 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e966d3c5-b228-4842-a3d6-9d9c9de192ce-utilities\") pod \"certified-operators-mr2xt\" (UID: \"e966d3c5-b228-4842-a3d6-9d9c9de192ce\") " pod="openshift-marketplace/certified-operators-mr2xt" Dec 03 16:17:14 crc kubenswrapper[4998]: I1203 16:17:14.561832 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e966d3c5-b228-4842-a3d6-9d9c9de192ce-utilities\") pod \"certified-operators-mr2xt\" (UID: \"e966d3c5-b228-4842-a3d6-9d9c9de192ce\") " pod="openshift-marketplace/certified-operators-mr2xt" Dec 03 16:17:14 crc kubenswrapper[4998]: I1203 16:17:14.562202 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e966d3c5-b228-4842-a3d6-9d9c9de192ce-catalog-content\") pod \"certified-operators-mr2xt\" (UID: \"e966d3c5-b228-4842-a3d6-9d9c9de192ce\") " pod="openshift-marketplace/certified-operators-mr2xt" Dec 03 16:17:14 crc kubenswrapper[4998]: I1203 16:17:14.584741 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbzdn\" (UniqueName: \"kubernetes.io/projected/e966d3c5-b228-4842-a3d6-9d9c9de192ce-kube-api-access-zbzdn\") pod \"certified-operators-mr2xt\" (UID: \"e966d3c5-b228-4842-a3d6-9d9c9de192ce\") " pod="openshift-marketplace/certified-operators-mr2xt" Dec 03 16:17:14 crc kubenswrapper[4998]: I1203 16:17:14.629926 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mr2xt" Dec 03 16:17:15 crc kubenswrapper[4998]: I1203 16:17:15.406624 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mr2xt"] Dec 03 16:17:15 crc kubenswrapper[4998]: W1203 16:17:15.412065 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode966d3c5_b228_4842_a3d6_9d9c9de192ce.slice/crio-90ff03526d8dc4e0977ec045293a1d2d7c65844abe7aab0f473fe190587b398d WatchSource:0}: Error finding container 90ff03526d8dc4e0977ec045293a1d2d7c65844abe7aab0f473fe190587b398d: Status 404 returned error can't find the container with id 90ff03526d8dc4e0977ec045293a1d2d7c65844abe7aab0f473fe190587b398d Dec 03 16:17:16 crc kubenswrapper[4998]: I1203 16:17:16.205792 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-n9fkd" event={"ID":"c74a341f-037c-472d-af3c-d76406b0f422","Type":"ContainerStarted","Data":"be510b60258edc12c8bf209ee1c0209e42809b7a9950c48f804e8a7952c72b2e"} Dec 03 16:17:16 crc kubenswrapper[4998]: I1203 16:17:16.207564 4998 generic.go:334] "Generic (PLEG): container finished" podID="e966d3c5-b228-4842-a3d6-9d9c9de192ce" containerID="a18202c06d065061206f07f77e7d40193a08515a657c4cf2d3c8272def46c4c5" exitCode=0 Dec 03 16:17:16 crc kubenswrapper[4998]: I1203 16:17:16.207628 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mr2xt" event={"ID":"e966d3c5-b228-4842-a3d6-9d9c9de192ce","Type":"ContainerDied","Data":"a18202c06d065061206f07f77e7d40193a08515a657c4cf2d3c8272def46c4c5"} Dec 03 16:17:16 crc kubenswrapper[4998]: I1203 16:17:16.207667 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mr2xt" event={"ID":"e966d3c5-b228-4842-a3d6-9d9c9de192ce","Type":"ContainerStarted","Data":"90ff03526d8dc4e0977ec045293a1d2d7c65844abe7aab0f473fe190587b398d"} Dec 03 16:17:16 crc kubenswrapper[4998]: I1203 16:17:16.227536 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-n9fkd" podStartSLOduration=1.759448538 podStartE2EDuration="7.227512003s" podCreationTimestamp="2025-12-03 16:17:09 +0000 UTC" firstStartedPulling="2025-12-03 16:17:09.823230106 +0000 UTC m=+808.434930329" lastFinishedPulling="2025-12-03 16:17:15.291293541 +0000 UTC m=+813.902993794" observedRunningTime="2025-12-03 16:17:16.223800107 +0000 UTC m=+814.835500330" watchObservedRunningTime="2025-12-03 16:17:16.227512003 +0000 UTC m=+814.839212226" Dec 03 16:17:18 crc kubenswrapper[4998]: I1203 16:17:18.224261 4998 generic.go:334] "Generic (PLEG): container finished" podID="e966d3c5-b228-4842-a3d6-9d9c9de192ce" containerID="4cb92f7c036bf0179f7922bf3af8a5b0f55d4433a014d6cc9893a94bf7e25659" exitCode=0 Dec 03 16:17:18 crc kubenswrapper[4998]: I1203 16:17:18.224338 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mr2xt" event={"ID":"e966d3c5-b228-4842-a3d6-9d9c9de192ce","Type":"ContainerDied","Data":"4cb92f7c036bf0179f7922bf3af8a5b0f55d4433a014d6cc9893a94bf7e25659"} Dec 03 16:17:18 crc kubenswrapper[4998]: I1203 16:17:18.417598 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5bc7h" Dec 03 16:17:18 crc kubenswrapper[4998]: I1203 16:17:18.417672 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5bc7h" Dec 03 16:17:18 crc kubenswrapper[4998]: I1203 16:17:18.483483 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5bc7h" Dec 03 16:17:19 crc kubenswrapper[4998]: I1203 16:17:19.237535 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mr2xt" event={"ID":"e966d3c5-b228-4842-a3d6-9d9c9de192ce","Type":"ContainerStarted","Data":"a87a50fa8efc2a8a9c2af7c7efef0bc521adbdadc114fefb3da94395e112890a"} Dec 03 16:17:19 crc kubenswrapper[4998]: I1203 16:17:19.257964 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mr2xt" podStartSLOduration=2.861418755 podStartE2EDuration="5.257945643s" podCreationTimestamp="2025-12-03 16:17:14 +0000 UTC" firstStartedPulling="2025-12-03 16:17:16.209430757 +0000 UTC m=+814.821130990" lastFinishedPulling="2025-12-03 16:17:18.605957655 +0000 UTC m=+817.217657878" observedRunningTime="2025-12-03 16:17:19.256782353 +0000 UTC m=+817.868482576" watchObservedRunningTime="2025-12-03 16:17:19.257945643 +0000 UTC m=+817.869645866" Dec 03 16:17:19 crc kubenswrapper[4998]: I1203 16:17:19.286080 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5bc7h" Dec 03 16:17:19 crc kubenswrapper[4998]: I1203 16:17:19.490276 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-plkwj" Dec 03 16:17:19 crc kubenswrapper[4998]: I1203 16:17:19.764665 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:19 crc kubenswrapper[4998]: I1203 16:17:19.764730 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:19 crc kubenswrapper[4998]: I1203 16:17:19.770423 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:20 crc kubenswrapper[4998]: I1203 16:17:20.251666 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-789646f4f5-xq522" Dec 03 16:17:20 crc kubenswrapper[4998]: I1203 16:17:20.312012 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-kk8mb"] Dec 03 16:17:20 crc kubenswrapper[4998]: I1203 16:17:20.892335 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5bc7h"] Dec 03 16:17:21 crc kubenswrapper[4998]: I1203 16:17:21.250679 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5bc7h" podUID="32577841-0e29-442c-b7e3-492335ab2d4c" containerName="registry-server" containerID="cri-o://b3ebef1adeb796121bf8c68d788da8039e5276ff66a6a27a610b90eb8183e691" gracePeriod=2 Dec 03 16:17:22 crc kubenswrapper[4998]: I1203 16:17:22.139072 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5bc7h" Dec 03 16:17:22 crc kubenswrapper[4998]: I1203 16:17:22.259640 4998 generic.go:334] "Generic (PLEG): container finished" podID="32577841-0e29-442c-b7e3-492335ab2d4c" containerID="b3ebef1adeb796121bf8c68d788da8039e5276ff66a6a27a610b90eb8183e691" exitCode=0 Dec 03 16:17:22 crc kubenswrapper[4998]: I1203 16:17:22.259685 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5bc7h" event={"ID":"32577841-0e29-442c-b7e3-492335ab2d4c","Type":"ContainerDied","Data":"b3ebef1adeb796121bf8c68d788da8039e5276ff66a6a27a610b90eb8183e691"} Dec 03 16:17:22 crc kubenswrapper[4998]: I1203 16:17:22.259714 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5bc7h" event={"ID":"32577841-0e29-442c-b7e3-492335ab2d4c","Type":"ContainerDied","Data":"48b2677e89acb69569386b27c01efbda514b786d9a9474574d44e7ddd28184b3"} Dec 03 16:17:22 crc kubenswrapper[4998]: I1203 16:17:22.259733 4998 scope.go:117] "RemoveContainer" containerID="b3ebef1adeb796121bf8c68d788da8039e5276ff66a6a27a610b90eb8183e691" Dec 03 16:17:22 crc kubenswrapper[4998]: I1203 16:17:22.259689 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5bc7h" Dec 03 16:17:22 crc kubenswrapper[4998]: I1203 16:17:22.264391 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32577841-0e29-442c-b7e3-492335ab2d4c-catalog-content\") pod \"32577841-0e29-442c-b7e3-492335ab2d4c\" (UID: \"32577841-0e29-442c-b7e3-492335ab2d4c\") " Dec 03 16:17:22 crc kubenswrapper[4998]: I1203 16:17:22.264535 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32577841-0e29-442c-b7e3-492335ab2d4c-utilities\") pod \"32577841-0e29-442c-b7e3-492335ab2d4c\" (UID: \"32577841-0e29-442c-b7e3-492335ab2d4c\") " Dec 03 16:17:22 crc kubenswrapper[4998]: I1203 16:17:22.264684 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psw76\" (UniqueName: \"kubernetes.io/projected/32577841-0e29-442c-b7e3-492335ab2d4c-kube-api-access-psw76\") pod \"32577841-0e29-442c-b7e3-492335ab2d4c\" (UID: \"32577841-0e29-442c-b7e3-492335ab2d4c\") " Dec 03 16:17:22 crc kubenswrapper[4998]: I1203 16:17:22.265300 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32577841-0e29-442c-b7e3-492335ab2d4c-utilities" (OuterVolumeSpecName: "utilities") pod "32577841-0e29-442c-b7e3-492335ab2d4c" (UID: "32577841-0e29-442c-b7e3-492335ab2d4c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:17:22 crc kubenswrapper[4998]: I1203 16:17:22.266143 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32577841-0e29-442c-b7e3-492335ab2d4c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:17:22 crc kubenswrapper[4998]: I1203 16:17:22.270306 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32577841-0e29-442c-b7e3-492335ab2d4c-kube-api-access-psw76" (OuterVolumeSpecName: "kube-api-access-psw76") pod "32577841-0e29-442c-b7e3-492335ab2d4c" (UID: "32577841-0e29-442c-b7e3-492335ab2d4c"). InnerVolumeSpecName "kube-api-access-psw76". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:17:22 crc kubenswrapper[4998]: I1203 16:17:22.283830 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32577841-0e29-442c-b7e3-492335ab2d4c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "32577841-0e29-442c-b7e3-492335ab2d4c" (UID: "32577841-0e29-442c-b7e3-492335ab2d4c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:17:22 crc kubenswrapper[4998]: I1203 16:17:22.307924 4998 scope.go:117] "RemoveContainer" containerID="5aafef86b40f875aac0d5b7eff13fbae6abd320dda74a9be51beab331a9f936f" Dec 03 16:17:22 crc kubenswrapper[4998]: I1203 16:17:22.325859 4998 scope.go:117] "RemoveContainer" containerID="a37e82548404bde528367b0111941471fe8ade2d8eaa3c74f2564a9c00e83364" Dec 03 16:17:22 crc kubenswrapper[4998]: I1203 16:17:22.348251 4998 scope.go:117] "RemoveContainer" containerID="b3ebef1adeb796121bf8c68d788da8039e5276ff66a6a27a610b90eb8183e691" Dec 03 16:17:22 crc kubenswrapper[4998]: E1203 16:17:22.348693 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3ebef1adeb796121bf8c68d788da8039e5276ff66a6a27a610b90eb8183e691\": container with ID starting with b3ebef1adeb796121bf8c68d788da8039e5276ff66a6a27a610b90eb8183e691 not found: ID does not exist" containerID="b3ebef1adeb796121bf8c68d788da8039e5276ff66a6a27a610b90eb8183e691" Dec 03 16:17:22 crc kubenswrapper[4998]: I1203 16:17:22.348737 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3ebef1adeb796121bf8c68d788da8039e5276ff66a6a27a610b90eb8183e691"} err="failed to get container status \"b3ebef1adeb796121bf8c68d788da8039e5276ff66a6a27a610b90eb8183e691\": rpc error: code = NotFound desc = could not find container \"b3ebef1adeb796121bf8c68d788da8039e5276ff66a6a27a610b90eb8183e691\": container with ID starting with b3ebef1adeb796121bf8c68d788da8039e5276ff66a6a27a610b90eb8183e691 not found: ID does not exist" Dec 03 16:17:22 crc kubenswrapper[4998]: I1203 16:17:22.348778 4998 scope.go:117] "RemoveContainer" containerID="5aafef86b40f875aac0d5b7eff13fbae6abd320dda74a9be51beab331a9f936f" Dec 03 16:17:22 crc kubenswrapper[4998]: E1203 16:17:22.349047 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5aafef86b40f875aac0d5b7eff13fbae6abd320dda74a9be51beab331a9f936f\": container with ID starting with 5aafef86b40f875aac0d5b7eff13fbae6abd320dda74a9be51beab331a9f936f not found: ID does not exist" containerID="5aafef86b40f875aac0d5b7eff13fbae6abd320dda74a9be51beab331a9f936f" Dec 03 16:17:22 crc kubenswrapper[4998]: I1203 16:17:22.349076 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5aafef86b40f875aac0d5b7eff13fbae6abd320dda74a9be51beab331a9f936f"} err="failed to get container status \"5aafef86b40f875aac0d5b7eff13fbae6abd320dda74a9be51beab331a9f936f\": rpc error: code = NotFound desc = could not find container \"5aafef86b40f875aac0d5b7eff13fbae6abd320dda74a9be51beab331a9f936f\": container with ID starting with 5aafef86b40f875aac0d5b7eff13fbae6abd320dda74a9be51beab331a9f936f not found: ID does not exist" Dec 03 16:17:22 crc kubenswrapper[4998]: I1203 16:17:22.349094 4998 scope.go:117] "RemoveContainer" containerID="a37e82548404bde528367b0111941471fe8ade2d8eaa3c74f2564a9c00e83364" Dec 03 16:17:22 crc kubenswrapper[4998]: E1203 16:17:22.349401 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a37e82548404bde528367b0111941471fe8ade2d8eaa3c74f2564a9c00e83364\": container with ID starting with a37e82548404bde528367b0111941471fe8ade2d8eaa3c74f2564a9c00e83364 not found: ID does not exist" containerID="a37e82548404bde528367b0111941471fe8ade2d8eaa3c74f2564a9c00e83364" Dec 03 16:17:22 crc kubenswrapper[4998]: I1203 16:17:22.349427 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a37e82548404bde528367b0111941471fe8ade2d8eaa3c74f2564a9c00e83364"} err="failed to get container status \"a37e82548404bde528367b0111941471fe8ade2d8eaa3c74f2564a9c00e83364\": rpc error: code = NotFound desc = could not find container \"a37e82548404bde528367b0111941471fe8ade2d8eaa3c74f2564a9c00e83364\": container with ID starting with a37e82548404bde528367b0111941471fe8ade2d8eaa3c74f2564a9c00e83364 not found: ID does not exist" Dec 03 16:17:22 crc kubenswrapper[4998]: I1203 16:17:22.367558 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32577841-0e29-442c-b7e3-492335ab2d4c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:17:22 crc kubenswrapper[4998]: I1203 16:17:22.367580 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-psw76\" (UniqueName: \"kubernetes.io/projected/32577841-0e29-442c-b7e3-492335ab2d4c-kube-api-access-psw76\") on node \"crc\" DevicePath \"\"" Dec 03 16:17:22 crc kubenswrapper[4998]: I1203 16:17:22.610103 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5bc7h"] Dec 03 16:17:22 crc kubenswrapper[4998]: I1203 16:17:22.617296 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5bc7h"] Dec 03 16:17:22 crc kubenswrapper[4998]: E1203 16:17:22.699538 4998 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32577841_0e29_442c_b7e3_492335ab2d4c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32577841_0e29_442c_b7e3_492335ab2d4c.slice/crio-48b2677e89acb69569386b27c01efbda514b786d9a9474574d44e7ddd28184b3\": RecentStats: unable to find data in memory cache]" Dec 03 16:17:23 crc kubenswrapper[4998]: I1203 16:17:23.687181 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32577841-0e29-442c-b7e3-492335ab2d4c" path="/var/lib/kubelet/pods/32577841-0e29-442c-b7e3-492335ab2d4c/volumes" Dec 03 16:17:24 crc kubenswrapper[4998]: I1203 16:17:24.630720 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mr2xt" Dec 03 16:17:24 crc kubenswrapper[4998]: I1203 16:17:24.631218 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mr2xt" Dec 03 16:17:24 crc kubenswrapper[4998]: I1203 16:17:24.704202 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mr2xt" Dec 03 16:17:25 crc kubenswrapper[4998]: I1203 16:17:25.344077 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mr2xt" Dec 03 16:17:26 crc kubenswrapper[4998]: I1203 16:17:26.486274 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mr2xt"] Dec 03 16:17:27 crc kubenswrapper[4998]: I1203 16:17:27.300940 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mr2xt" podUID="e966d3c5-b228-4842-a3d6-9d9c9de192ce" containerName="registry-server" containerID="cri-o://a87a50fa8efc2a8a9c2af7c7efef0bc521adbdadc114fefb3da94395e112890a" gracePeriod=2 Dec 03 16:17:28 crc kubenswrapper[4998]: I1203 16:17:28.310068 4998 generic.go:334] "Generic (PLEG): container finished" podID="e966d3c5-b228-4842-a3d6-9d9c9de192ce" containerID="a87a50fa8efc2a8a9c2af7c7efef0bc521adbdadc114fefb3da94395e112890a" exitCode=0 Dec 03 16:17:28 crc kubenswrapper[4998]: I1203 16:17:28.310133 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mr2xt" event={"ID":"e966d3c5-b228-4842-a3d6-9d9c9de192ce","Type":"ContainerDied","Data":"a87a50fa8efc2a8a9c2af7c7efef0bc521adbdadc114fefb3da94395e112890a"} Dec 03 16:17:28 crc kubenswrapper[4998]: I1203 16:17:28.484638 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mr2xt" Dec 03 16:17:28 crc kubenswrapper[4998]: I1203 16:17:28.553827 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbzdn\" (UniqueName: \"kubernetes.io/projected/e966d3c5-b228-4842-a3d6-9d9c9de192ce-kube-api-access-zbzdn\") pod \"e966d3c5-b228-4842-a3d6-9d9c9de192ce\" (UID: \"e966d3c5-b228-4842-a3d6-9d9c9de192ce\") " Dec 03 16:17:28 crc kubenswrapper[4998]: I1203 16:17:28.553888 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e966d3c5-b228-4842-a3d6-9d9c9de192ce-utilities\") pod \"e966d3c5-b228-4842-a3d6-9d9c9de192ce\" (UID: \"e966d3c5-b228-4842-a3d6-9d9c9de192ce\") " Dec 03 16:17:28 crc kubenswrapper[4998]: I1203 16:17:28.553998 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e966d3c5-b228-4842-a3d6-9d9c9de192ce-catalog-content\") pod \"e966d3c5-b228-4842-a3d6-9d9c9de192ce\" (UID: \"e966d3c5-b228-4842-a3d6-9d9c9de192ce\") " Dec 03 16:17:28 crc kubenswrapper[4998]: I1203 16:17:28.554913 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e966d3c5-b228-4842-a3d6-9d9c9de192ce-utilities" (OuterVolumeSpecName: "utilities") pod "e966d3c5-b228-4842-a3d6-9d9c9de192ce" (UID: "e966d3c5-b228-4842-a3d6-9d9c9de192ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:17:28 crc kubenswrapper[4998]: I1203 16:17:28.560941 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e966d3c5-b228-4842-a3d6-9d9c9de192ce-kube-api-access-zbzdn" (OuterVolumeSpecName: "kube-api-access-zbzdn") pod "e966d3c5-b228-4842-a3d6-9d9c9de192ce" (UID: "e966d3c5-b228-4842-a3d6-9d9c9de192ce"). InnerVolumeSpecName "kube-api-access-zbzdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:17:28 crc kubenswrapper[4998]: I1203 16:17:28.601370 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e966d3c5-b228-4842-a3d6-9d9c9de192ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e966d3c5-b228-4842-a3d6-9d9c9de192ce" (UID: "e966d3c5-b228-4842-a3d6-9d9c9de192ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:17:28 crc kubenswrapper[4998]: I1203 16:17:28.655834 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e966d3c5-b228-4842-a3d6-9d9c9de192ce-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:17:28 crc kubenswrapper[4998]: I1203 16:17:28.655883 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbzdn\" (UniqueName: \"kubernetes.io/projected/e966d3c5-b228-4842-a3d6-9d9c9de192ce-kube-api-access-zbzdn\") on node \"crc\" DevicePath \"\"" Dec 03 16:17:28 crc kubenswrapper[4998]: I1203 16:17:28.655897 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e966d3c5-b228-4842-a3d6-9d9c9de192ce-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:17:29 crc kubenswrapper[4998]: I1203 16:17:29.318799 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mr2xt" event={"ID":"e966d3c5-b228-4842-a3d6-9d9c9de192ce","Type":"ContainerDied","Data":"90ff03526d8dc4e0977ec045293a1d2d7c65844abe7aab0f473fe190587b398d"} Dec 03 16:17:29 crc kubenswrapper[4998]: I1203 16:17:29.319113 4998 scope.go:117] "RemoveContainer" containerID="a87a50fa8efc2a8a9c2af7c7efef0bc521adbdadc114fefb3da94395e112890a" Dec 03 16:17:29 crc kubenswrapper[4998]: I1203 16:17:29.318882 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mr2xt" Dec 03 16:17:29 crc kubenswrapper[4998]: I1203 16:17:29.336972 4998 scope.go:117] "RemoveContainer" containerID="4cb92f7c036bf0179f7922bf3af8a5b0f55d4433a014d6cc9893a94bf7e25659" Dec 03 16:17:29 crc kubenswrapper[4998]: I1203 16:17:29.350195 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mr2xt"] Dec 03 16:17:29 crc kubenswrapper[4998]: I1203 16:17:29.355015 4998 scope.go:117] "RemoveContainer" containerID="a18202c06d065061206f07f77e7d40193a08515a657c4cf2d3c8272def46c4c5" Dec 03 16:17:29 crc kubenswrapper[4998]: I1203 16:17:29.357842 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mr2xt"] Dec 03 16:17:29 crc kubenswrapper[4998]: I1203 16:17:29.464899 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7rcw5" Dec 03 16:17:29 crc kubenswrapper[4998]: I1203 16:17:29.686365 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e966d3c5-b228-4842-a3d6-9d9c9de192ce" path="/var/lib/kubelet/pods/e966d3c5-b228-4842-a3d6-9d9c9de192ce/volumes" Dec 03 16:17:44 crc kubenswrapper[4998]: I1203 16:17:44.327584 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl"] Dec 03 16:17:44 crc kubenswrapper[4998]: E1203 16:17:44.328423 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32577841-0e29-442c-b7e3-492335ab2d4c" containerName="extract-content" Dec 03 16:17:44 crc kubenswrapper[4998]: I1203 16:17:44.328445 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="32577841-0e29-442c-b7e3-492335ab2d4c" containerName="extract-content" Dec 03 16:17:44 crc kubenswrapper[4998]: E1203 16:17:44.328461 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e966d3c5-b228-4842-a3d6-9d9c9de192ce" containerName="extract-utilities" Dec 03 16:17:44 crc kubenswrapper[4998]: I1203 16:17:44.328470 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="e966d3c5-b228-4842-a3d6-9d9c9de192ce" containerName="extract-utilities" Dec 03 16:17:44 crc kubenswrapper[4998]: E1203 16:17:44.328485 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32577841-0e29-442c-b7e3-492335ab2d4c" containerName="extract-utilities" Dec 03 16:17:44 crc kubenswrapper[4998]: I1203 16:17:44.328493 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="32577841-0e29-442c-b7e3-492335ab2d4c" containerName="extract-utilities" Dec 03 16:17:44 crc kubenswrapper[4998]: E1203 16:17:44.328511 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e966d3c5-b228-4842-a3d6-9d9c9de192ce" containerName="extract-content" Dec 03 16:17:44 crc kubenswrapper[4998]: I1203 16:17:44.328519 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="e966d3c5-b228-4842-a3d6-9d9c9de192ce" containerName="extract-content" Dec 03 16:17:44 crc kubenswrapper[4998]: E1203 16:17:44.328531 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32577841-0e29-442c-b7e3-492335ab2d4c" containerName="registry-server" Dec 03 16:17:44 crc kubenswrapper[4998]: I1203 16:17:44.328541 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="32577841-0e29-442c-b7e3-492335ab2d4c" containerName="registry-server" Dec 03 16:17:44 crc kubenswrapper[4998]: E1203 16:17:44.328554 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e966d3c5-b228-4842-a3d6-9d9c9de192ce" containerName="registry-server" Dec 03 16:17:44 crc kubenswrapper[4998]: I1203 16:17:44.328563 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="e966d3c5-b228-4842-a3d6-9d9c9de192ce" containerName="registry-server" Dec 03 16:17:44 crc kubenswrapper[4998]: I1203 16:17:44.328697 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="32577841-0e29-442c-b7e3-492335ab2d4c" containerName="registry-server" Dec 03 16:17:44 crc kubenswrapper[4998]: I1203 16:17:44.328714 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="e966d3c5-b228-4842-a3d6-9d9c9de192ce" containerName="registry-server" Dec 03 16:17:44 crc kubenswrapper[4998]: I1203 16:17:44.329985 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl" Dec 03 16:17:44 crc kubenswrapper[4998]: I1203 16:17:44.335445 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 16:17:44 crc kubenswrapper[4998]: I1203 16:17:44.335549 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl"] Dec 03 16:17:44 crc kubenswrapper[4998]: I1203 16:17:44.473374 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8xdk\" (UniqueName: \"kubernetes.io/projected/f399c383-c804-4c6e-a247-0c09ca313644-kube-api-access-c8xdk\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl\" (UID: \"f399c383-c804-4c6e-a247-0c09ca313644\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl" Dec 03 16:17:44 crc kubenswrapper[4998]: I1203 16:17:44.473432 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f399c383-c804-4c6e-a247-0c09ca313644-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl\" (UID: \"f399c383-c804-4c6e-a247-0c09ca313644\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl" Dec 03 16:17:44 crc kubenswrapper[4998]: I1203 16:17:44.473465 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f399c383-c804-4c6e-a247-0c09ca313644-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl\" (UID: \"f399c383-c804-4c6e-a247-0c09ca313644\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl" Dec 03 16:17:44 crc kubenswrapper[4998]: I1203 16:17:44.574627 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8xdk\" (UniqueName: \"kubernetes.io/projected/f399c383-c804-4c6e-a247-0c09ca313644-kube-api-access-c8xdk\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl\" (UID: \"f399c383-c804-4c6e-a247-0c09ca313644\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl" Dec 03 16:17:44 crc kubenswrapper[4998]: I1203 16:17:44.574680 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f399c383-c804-4c6e-a247-0c09ca313644-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl\" (UID: \"f399c383-c804-4c6e-a247-0c09ca313644\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl" Dec 03 16:17:44 crc kubenswrapper[4998]: I1203 16:17:44.574701 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f399c383-c804-4c6e-a247-0c09ca313644-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl\" (UID: \"f399c383-c804-4c6e-a247-0c09ca313644\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl" Dec 03 16:17:44 crc kubenswrapper[4998]: I1203 16:17:44.575267 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f399c383-c804-4c6e-a247-0c09ca313644-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl\" (UID: \"f399c383-c804-4c6e-a247-0c09ca313644\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl" Dec 03 16:17:44 crc kubenswrapper[4998]: I1203 16:17:44.575312 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f399c383-c804-4c6e-a247-0c09ca313644-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl\" (UID: \"f399c383-c804-4c6e-a247-0c09ca313644\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl" Dec 03 16:17:44 crc kubenswrapper[4998]: I1203 16:17:44.591421 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8xdk\" (UniqueName: \"kubernetes.io/projected/f399c383-c804-4c6e-a247-0c09ca313644-kube-api-access-c8xdk\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl\" (UID: \"f399c383-c804-4c6e-a247-0c09ca313644\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl" Dec 03 16:17:44 crc kubenswrapper[4998]: I1203 16:17:44.646120 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl" Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.032703 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl"] Dec 03 16:17:45 crc kubenswrapper[4998]: W1203 16:17:45.040851 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf399c383_c804_4c6e_a247_0c09ca313644.slice/crio-627ab9ac788b16369bc1d11b1b20c8e174a0e1335e0dc9f338dd950f3c62f204 WatchSource:0}: Error finding container 627ab9ac788b16369bc1d11b1b20c8e174a0e1335e0dc9f338dd950f3c62f204: Status 404 returned error can't find the container with id 627ab9ac788b16369bc1d11b1b20c8e174a0e1335e0dc9f338dd950f3c62f204 Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.365857 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-kk8mb" podUID="03021110-7732-4dff-a5e3-f481e8b7c0d6" containerName="console" containerID="cri-o://86442755efbc807979332830ff8cf6a8cdce2a791eb5a81ea3f7aac061669b6d" gracePeriod=15 Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.417160 4998 generic.go:334] "Generic (PLEG): container finished" podID="f399c383-c804-4c6e-a247-0c09ca313644" containerID="2fb9a4f6c46138908ef68c4ae5bf5639d056dfaa58565e82125d22bcda559fc3" exitCode=0 Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.417209 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl" event={"ID":"f399c383-c804-4c6e-a247-0c09ca313644","Type":"ContainerDied","Data":"2fb9a4f6c46138908ef68c4ae5bf5639d056dfaa58565e82125d22bcda559fc3"} Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.417258 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl" event={"ID":"f399c383-c804-4c6e-a247-0c09ca313644","Type":"ContainerStarted","Data":"627ab9ac788b16369bc1d11b1b20c8e174a0e1335e0dc9f338dd950f3c62f204"} Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.697957 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-kk8mb_03021110-7732-4dff-a5e3-f481e8b7c0d6/console/0.log" Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.698032 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.789794 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjd2p\" (UniqueName: \"kubernetes.io/projected/03021110-7732-4dff-a5e3-f481e8b7c0d6-kube-api-access-pjd2p\") pod \"03021110-7732-4dff-a5e3-f481e8b7c0d6\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.789861 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03021110-7732-4dff-a5e3-f481e8b7c0d6-trusted-ca-bundle\") pod \"03021110-7732-4dff-a5e3-f481e8b7c0d6\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.789952 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/03021110-7732-4dff-a5e3-f481e8b7c0d6-service-ca\") pod \"03021110-7732-4dff-a5e3-f481e8b7c0d6\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.789984 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/03021110-7732-4dff-a5e3-f481e8b7c0d6-console-oauth-config\") pod \"03021110-7732-4dff-a5e3-f481e8b7c0d6\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.790017 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/03021110-7732-4dff-a5e3-f481e8b7c0d6-console-config\") pod \"03021110-7732-4dff-a5e3-f481e8b7c0d6\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.790047 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/03021110-7732-4dff-a5e3-f481e8b7c0d6-console-serving-cert\") pod \"03021110-7732-4dff-a5e3-f481e8b7c0d6\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.790075 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/03021110-7732-4dff-a5e3-f481e8b7c0d6-oauth-serving-cert\") pod \"03021110-7732-4dff-a5e3-f481e8b7c0d6\" (UID: \"03021110-7732-4dff-a5e3-f481e8b7c0d6\") " Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.790711 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03021110-7732-4dff-a5e3-f481e8b7c0d6-service-ca" (OuterVolumeSpecName: "service-ca") pod "03021110-7732-4dff-a5e3-f481e8b7c0d6" (UID: "03021110-7732-4dff-a5e3-f481e8b7c0d6"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.791157 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03021110-7732-4dff-a5e3-f481e8b7c0d6-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "03021110-7732-4dff-a5e3-f481e8b7c0d6" (UID: "03021110-7732-4dff-a5e3-f481e8b7c0d6"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.791645 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03021110-7732-4dff-a5e3-f481e8b7c0d6-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "03021110-7732-4dff-a5e3-f481e8b7c0d6" (UID: "03021110-7732-4dff-a5e3-f481e8b7c0d6"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.791773 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03021110-7732-4dff-a5e3-f481e8b7c0d6-console-config" (OuterVolumeSpecName: "console-config") pod "03021110-7732-4dff-a5e3-f481e8b7c0d6" (UID: "03021110-7732-4dff-a5e3-f481e8b7c0d6"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.796294 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03021110-7732-4dff-a5e3-f481e8b7c0d6-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "03021110-7732-4dff-a5e3-f481e8b7c0d6" (UID: "03021110-7732-4dff-a5e3-f481e8b7c0d6"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.796633 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03021110-7732-4dff-a5e3-f481e8b7c0d6-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "03021110-7732-4dff-a5e3-f481e8b7c0d6" (UID: "03021110-7732-4dff-a5e3-f481e8b7c0d6"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.797215 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03021110-7732-4dff-a5e3-f481e8b7c0d6-kube-api-access-pjd2p" (OuterVolumeSpecName: "kube-api-access-pjd2p") pod "03021110-7732-4dff-a5e3-f481e8b7c0d6" (UID: "03021110-7732-4dff-a5e3-f481e8b7c0d6"). InnerVolumeSpecName "kube-api-access-pjd2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.891813 4998 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/03021110-7732-4dff-a5e3-f481e8b7c0d6-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.891869 4998 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/03021110-7732-4dff-a5e3-f481e8b7c0d6-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.891882 4998 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/03021110-7732-4dff-a5e3-f481e8b7c0d6-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.891893 4998 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/03021110-7732-4dff-a5e3-f481e8b7c0d6-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.891904 4998 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/03021110-7732-4dff-a5e3-f481e8b7c0d6-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.891914 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjd2p\" (UniqueName: \"kubernetes.io/projected/03021110-7732-4dff-a5e3-f481e8b7c0d6-kube-api-access-pjd2p\") on node \"crc\" DevicePath \"\"" Dec 03 16:17:45 crc kubenswrapper[4998]: I1203 16:17:45.891927 4998 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03021110-7732-4dff-a5e3-f481e8b7c0d6-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:17:46 crc kubenswrapper[4998]: I1203 16:17:46.426014 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-kk8mb_03021110-7732-4dff-a5e3-f481e8b7c0d6/console/0.log" Dec 03 16:17:46 crc kubenswrapper[4998]: I1203 16:17:46.426064 4998 generic.go:334] "Generic (PLEG): container finished" podID="03021110-7732-4dff-a5e3-f481e8b7c0d6" containerID="86442755efbc807979332830ff8cf6a8cdce2a791eb5a81ea3f7aac061669b6d" exitCode=2 Dec 03 16:17:46 crc kubenswrapper[4998]: I1203 16:17:46.426095 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kk8mb" event={"ID":"03021110-7732-4dff-a5e3-f481e8b7c0d6","Type":"ContainerDied","Data":"86442755efbc807979332830ff8cf6a8cdce2a791eb5a81ea3f7aac061669b6d"} Dec 03 16:17:46 crc kubenswrapper[4998]: I1203 16:17:46.426120 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kk8mb" event={"ID":"03021110-7732-4dff-a5e3-f481e8b7c0d6","Type":"ContainerDied","Data":"ba405cb10a18a7f717e33aa3807d4a2d514dc108ff7c5e11d02a6883fa993d10"} Dec 03 16:17:46 crc kubenswrapper[4998]: I1203 16:17:46.426136 4998 scope.go:117] "RemoveContainer" containerID="86442755efbc807979332830ff8cf6a8cdce2a791eb5a81ea3f7aac061669b6d" Dec 03 16:17:46 crc kubenswrapper[4998]: I1203 16:17:46.426257 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kk8mb" Dec 03 16:17:46 crc kubenswrapper[4998]: I1203 16:17:46.455895 4998 scope.go:117] "RemoveContainer" containerID="86442755efbc807979332830ff8cf6a8cdce2a791eb5a81ea3f7aac061669b6d" Dec 03 16:17:46 crc kubenswrapper[4998]: E1203 16:17:46.456482 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86442755efbc807979332830ff8cf6a8cdce2a791eb5a81ea3f7aac061669b6d\": container with ID starting with 86442755efbc807979332830ff8cf6a8cdce2a791eb5a81ea3f7aac061669b6d not found: ID does not exist" containerID="86442755efbc807979332830ff8cf6a8cdce2a791eb5a81ea3f7aac061669b6d" Dec 03 16:17:46 crc kubenswrapper[4998]: I1203 16:17:46.456545 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86442755efbc807979332830ff8cf6a8cdce2a791eb5a81ea3f7aac061669b6d"} err="failed to get container status \"86442755efbc807979332830ff8cf6a8cdce2a791eb5a81ea3f7aac061669b6d\": rpc error: code = NotFound desc = could not find container \"86442755efbc807979332830ff8cf6a8cdce2a791eb5a81ea3f7aac061669b6d\": container with ID starting with 86442755efbc807979332830ff8cf6a8cdce2a791eb5a81ea3f7aac061669b6d not found: ID does not exist" Dec 03 16:17:46 crc kubenswrapper[4998]: I1203 16:17:46.458627 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-kk8mb"] Dec 03 16:17:46 crc kubenswrapper[4998]: I1203 16:17:46.462450 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-kk8mb"] Dec 03 16:17:47 crc kubenswrapper[4998]: I1203 16:17:47.438106 4998 generic.go:334] "Generic (PLEG): container finished" podID="f399c383-c804-4c6e-a247-0c09ca313644" containerID="ba66b374333a5fdb72ccf6d5e94100b72a5ea3b8f009a1dc2209f7aeccd04235" exitCode=0 Dec 03 16:17:47 crc kubenswrapper[4998]: I1203 16:17:47.438217 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl" event={"ID":"f399c383-c804-4c6e-a247-0c09ca313644","Type":"ContainerDied","Data":"ba66b374333a5fdb72ccf6d5e94100b72a5ea3b8f009a1dc2209f7aeccd04235"} Dec 03 16:17:47 crc kubenswrapper[4998]: I1203 16:17:47.688977 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03021110-7732-4dff-a5e3-f481e8b7c0d6" path="/var/lib/kubelet/pods/03021110-7732-4dff-a5e3-f481e8b7c0d6/volumes" Dec 03 16:17:48 crc kubenswrapper[4998]: I1203 16:17:48.452400 4998 generic.go:334] "Generic (PLEG): container finished" podID="f399c383-c804-4c6e-a247-0c09ca313644" containerID="5090841dbffa482a48a981b787b8fd00acfca8c18e9def5fb32bf36be5c397a1" exitCode=0 Dec 03 16:17:48 crc kubenswrapper[4998]: I1203 16:17:48.452462 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl" event={"ID":"f399c383-c804-4c6e-a247-0c09ca313644","Type":"ContainerDied","Data":"5090841dbffa482a48a981b787b8fd00acfca8c18e9def5fb32bf36be5c397a1"} Dec 03 16:17:49 crc kubenswrapper[4998]: I1203 16:17:49.740880 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl" Dec 03 16:17:49 crc kubenswrapper[4998]: I1203 16:17:49.844711 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8xdk\" (UniqueName: \"kubernetes.io/projected/f399c383-c804-4c6e-a247-0c09ca313644-kube-api-access-c8xdk\") pod \"f399c383-c804-4c6e-a247-0c09ca313644\" (UID: \"f399c383-c804-4c6e-a247-0c09ca313644\") " Dec 03 16:17:49 crc kubenswrapper[4998]: I1203 16:17:49.844879 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f399c383-c804-4c6e-a247-0c09ca313644-bundle\") pod \"f399c383-c804-4c6e-a247-0c09ca313644\" (UID: \"f399c383-c804-4c6e-a247-0c09ca313644\") " Dec 03 16:17:49 crc kubenswrapper[4998]: I1203 16:17:49.844920 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f399c383-c804-4c6e-a247-0c09ca313644-util\") pod \"f399c383-c804-4c6e-a247-0c09ca313644\" (UID: \"f399c383-c804-4c6e-a247-0c09ca313644\") " Dec 03 16:17:49 crc kubenswrapper[4998]: I1203 16:17:49.846878 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f399c383-c804-4c6e-a247-0c09ca313644-bundle" (OuterVolumeSpecName: "bundle") pod "f399c383-c804-4c6e-a247-0c09ca313644" (UID: "f399c383-c804-4c6e-a247-0c09ca313644"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:17:49 crc kubenswrapper[4998]: I1203 16:17:49.857993 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f399c383-c804-4c6e-a247-0c09ca313644-util" (OuterVolumeSpecName: "util") pod "f399c383-c804-4c6e-a247-0c09ca313644" (UID: "f399c383-c804-4c6e-a247-0c09ca313644"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:17:49 crc kubenswrapper[4998]: I1203 16:17:49.858040 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f399c383-c804-4c6e-a247-0c09ca313644-kube-api-access-c8xdk" (OuterVolumeSpecName: "kube-api-access-c8xdk") pod "f399c383-c804-4c6e-a247-0c09ca313644" (UID: "f399c383-c804-4c6e-a247-0c09ca313644"). InnerVolumeSpecName "kube-api-access-c8xdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:17:49 crc kubenswrapper[4998]: I1203 16:17:49.946213 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8xdk\" (UniqueName: \"kubernetes.io/projected/f399c383-c804-4c6e-a247-0c09ca313644-kube-api-access-c8xdk\") on node \"crc\" DevicePath \"\"" Dec 03 16:17:49 crc kubenswrapper[4998]: I1203 16:17:49.946276 4998 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f399c383-c804-4c6e-a247-0c09ca313644-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:17:49 crc kubenswrapper[4998]: I1203 16:17:49.946297 4998 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f399c383-c804-4c6e-a247-0c09ca313644-util\") on node \"crc\" DevicePath \"\"" Dec 03 16:17:50 crc kubenswrapper[4998]: I1203 16:17:50.470449 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl" event={"ID":"f399c383-c804-4c6e-a247-0c09ca313644","Type":"ContainerDied","Data":"627ab9ac788b16369bc1d11b1b20c8e174a0e1335e0dc9f338dd950f3c62f204"} Dec 03 16:17:50 crc kubenswrapper[4998]: I1203 16:17:50.470498 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="627ab9ac788b16369bc1d11b1b20c8e174a0e1335e0dc9f338dd950f3c62f204" Dec 03 16:17:50 crc kubenswrapper[4998]: I1203 16:17:50.470552 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl" Dec 03 16:17:50 crc kubenswrapper[4998]: I1203 16:17:50.486499 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-z9mwr"] Dec 03 16:17:50 crc kubenswrapper[4998]: E1203 16:17:50.486776 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03021110-7732-4dff-a5e3-f481e8b7c0d6" containerName="console" Dec 03 16:17:50 crc kubenswrapper[4998]: I1203 16:17:50.486791 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="03021110-7732-4dff-a5e3-f481e8b7c0d6" containerName="console" Dec 03 16:17:50 crc kubenswrapper[4998]: E1203 16:17:50.486813 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f399c383-c804-4c6e-a247-0c09ca313644" containerName="extract" Dec 03 16:17:50 crc kubenswrapper[4998]: I1203 16:17:50.486820 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="f399c383-c804-4c6e-a247-0c09ca313644" containerName="extract" Dec 03 16:17:50 crc kubenswrapper[4998]: E1203 16:17:50.486834 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f399c383-c804-4c6e-a247-0c09ca313644" containerName="pull" Dec 03 16:17:50 crc kubenswrapper[4998]: I1203 16:17:50.486840 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="f399c383-c804-4c6e-a247-0c09ca313644" containerName="pull" Dec 03 16:17:50 crc kubenswrapper[4998]: E1203 16:17:50.486849 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f399c383-c804-4c6e-a247-0c09ca313644" containerName="util" Dec 03 16:17:50 crc kubenswrapper[4998]: I1203 16:17:50.486855 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="f399c383-c804-4c6e-a247-0c09ca313644" containerName="util" Dec 03 16:17:50 crc kubenswrapper[4998]: I1203 16:17:50.486970 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="03021110-7732-4dff-a5e3-f481e8b7c0d6" containerName="console" Dec 03 16:17:50 crc kubenswrapper[4998]: I1203 16:17:50.486980 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="f399c383-c804-4c6e-a247-0c09ca313644" containerName="extract" Dec 03 16:17:50 crc kubenswrapper[4998]: I1203 16:17:50.488894 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z9mwr" Dec 03 16:17:50 crc kubenswrapper[4998]: I1203 16:17:50.503563 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z9mwr"] Dec 03 16:17:50 crc kubenswrapper[4998]: I1203 16:17:50.555541 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5c42d90-b0b9-4ac4-a5d9-d355a5d27388-catalog-content\") pod \"community-operators-z9mwr\" (UID: \"e5c42d90-b0b9-4ac4-a5d9-d355a5d27388\") " pod="openshift-marketplace/community-operators-z9mwr" Dec 03 16:17:50 crc kubenswrapper[4998]: I1203 16:17:50.555865 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5c42d90-b0b9-4ac4-a5d9-d355a5d27388-utilities\") pod \"community-operators-z9mwr\" (UID: \"e5c42d90-b0b9-4ac4-a5d9-d355a5d27388\") " pod="openshift-marketplace/community-operators-z9mwr" Dec 03 16:17:50 crc kubenswrapper[4998]: I1203 16:17:50.555918 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88rds\" (UniqueName: \"kubernetes.io/projected/e5c42d90-b0b9-4ac4-a5d9-d355a5d27388-kube-api-access-88rds\") pod \"community-operators-z9mwr\" (UID: \"e5c42d90-b0b9-4ac4-a5d9-d355a5d27388\") " pod="openshift-marketplace/community-operators-z9mwr" Dec 03 16:17:50 crc kubenswrapper[4998]: I1203 16:17:50.657421 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5c42d90-b0b9-4ac4-a5d9-d355a5d27388-catalog-content\") pod \"community-operators-z9mwr\" (UID: \"e5c42d90-b0b9-4ac4-a5d9-d355a5d27388\") " pod="openshift-marketplace/community-operators-z9mwr" Dec 03 16:17:50 crc kubenswrapper[4998]: I1203 16:17:50.657931 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5c42d90-b0b9-4ac4-a5d9-d355a5d27388-utilities\") pod \"community-operators-z9mwr\" (UID: \"e5c42d90-b0b9-4ac4-a5d9-d355a5d27388\") " pod="openshift-marketplace/community-operators-z9mwr" Dec 03 16:17:50 crc kubenswrapper[4998]: I1203 16:17:50.657970 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88rds\" (UniqueName: \"kubernetes.io/projected/e5c42d90-b0b9-4ac4-a5d9-d355a5d27388-kube-api-access-88rds\") pod \"community-operators-z9mwr\" (UID: \"e5c42d90-b0b9-4ac4-a5d9-d355a5d27388\") " pod="openshift-marketplace/community-operators-z9mwr" Dec 03 16:17:50 crc kubenswrapper[4998]: I1203 16:17:50.658006 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5c42d90-b0b9-4ac4-a5d9-d355a5d27388-catalog-content\") pod \"community-operators-z9mwr\" (UID: \"e5c42d90-b0b9-4ac4-a5d9-d355a5d27388\") " pod="openshift-marketplace/community-operators-z9mwr" Dec 03 16:17:50 crc kubenswrapper[4998]: I1203 16:17:50.658237 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5c42d90-b0b9-4ac4-a5d9-d355a5d27388-utilities\") pod \"community-operators-z9mwr\" (UID: \"e5c42d90-b0b9-4ac4-a5d9-d355a5d27388\") " pod="openshift-marketplace/community-operators-z9mwr" Dec 03 16:17:50 crc kubenswrapper[4998]: I1203 16:17:50.678796 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88rds\" (UniqueName: \"kubernetes.io/projected/e5c42d90-b0b9-4ac4-a5d9-d355a5d27388-kube-api-access-88rds\") pod \"community-operators-z9mwr\" (UID: \"e5c42d90-b0b9-4ac4-a5d9-d355a5d27388\") " pod="openshift-marketplace/community-operators-z9mwr" Dec 03 16:17:50 crc kubenswrapper[4998]: I1203 16:17:50.805300 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z9mwr" Dec 03 16:17:51 crc kubenswrapper[4998]: I1203 16:17:51.045684 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z9mwr"] Dec 03 16:17:51 crc kubenswrapper[4998]: I1203 16:17:51.478130 4998 generic.go:334] "Generic (PLEG): container finished" podID="e5c42d90-b0b9-4ac4-a5d9-d355a5d27388" containerID="09188b4a9d7e58c5a70bbefbf7122a0c00dd239bf19a8eece428f338763fa8f6" exitCode=0 Dec 03 16:17:51 crc kubenswrapper[4998]: I1203 16:17:51.478171 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z9mwr" event={"ID":"e5c42d90-b0b9-4ac4-a5d9-d355a5d27388","Type":"ContainerDied","Data":"09188b4a9d7e58c5a70bbefbf7122a0c00dd239bf19a8eece428f338763fa8f6"} Dec 03 16:17:51 crc kubenswrapper[4998]: I1203 16:17:51.478498 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z9mwr" event={"ID":"e5c42d90-b0b9-4ac4-a5d9-d355a5d27388","Type":"ContainerStarted","Data":"bb8aa35fd610e5ab7d9a397f9fe5eec674830e08367d94708834c2f6fae39734"} Dec 03 16:17:52 crc kubenswrapper[4998]: I1203 16:17:52.486969 4998 generic.go:334] "Generic (PLEG): container finished" podID="e5c42d90-b0b9-4ac4-a5d9-d355a5d27388" containerID="59febff8a36410df271852bdbf7fee5eef27e88a0711f2ce433a3bc1183aa7d8" exitCode=0 Dec 03 16:17:52 crc kubenswrapper[4998]: I1203 16:17:52.487056 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z9mwr" event={"ID":"e5c42d90-b0b9-4ac4-a5d9-d355a5d27388","Type":"ContainerDied","Data":"59febff8a36410df271852bdbf7fee5eef27e88a0711f2ce433a3bc1183aa7d8"} Dec 03 16:17:53 crc kubenswrapper[4998]: I1203 16:17:53.624858 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z9mwr" event={"ID":"e5c42d90-b0b9-4ac4-a5d9-d355a5d27388","Type":"ContainerStarted","Data":"c090b865321578dcd456e89110d9af6321784b65249827f432217c12a183fd4d"} Dec 03 16:17:53 crc kubenswrapper[4998]: I1203 16:17:53.650918 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-z9mwr" podStartSLOduration=2.239567298 podStartE2EDuration="3.650902394s" podCreationTimestamp="2025-12-03 16:17:50 +0000 UTC" firstStartedPulling="2025-12-03 16:17:51.479533483 +0000 UTC m=+850.091233706" lastFinishedPulling="2025-12-03 16:17:52.890868579 +0000 UTC m=+851.502568802" observedRunningTime="2025-12-03 16:17:53.648354968 +0000 UTC m=+852.260055191" watchObservedRunningTime="2025-12-03 16:17:53.650902394 +0000 UTC m=+852.262602617" Dec 03 16:17:57 crc kubenswrapper[4998]: I1203 16:17:57.985106 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5d9448b85b-kszzj"] Dec 03 16:17:57 crc kubenswrapper[4998]: I1203 16:17:57.985948 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5d9448b85b-kszzj" Dec 03 16:17:57 crc kubenswrapper[4998]: I1203 16:17:57.990938 4998 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 03 16:17:57 crc kubenswrapper[4998]: I1203 16:17:57.992497 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 03 16:17:57 crc kubenswrapper[4998]: I1203 16:17:57.993113 4998 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 03 16:17:57 crc kubenswrapper[4998]: I1203 16:17:57.993677 4998 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-bsbk8" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.003803 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5d9448b85b-kszzj"] Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.004511 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.048691 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dfqp\" (UniqueName: \"kubernetes.io/projected/a4f19703-b458-4f44-b824-6c249ccf6194-kube-api-access-4dfqp\") pod \"metallb-operator-controller-manager-5d9448b85b-kszzj\" (UID: \"a4f19703-b458-4f44-b824-6c249ccf6194\") " pod="metallb-system/metallb-operator-controller-manager-5d9448b85b-kszzj" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.048760 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a4f19703-b458-4f44-b824-6c249ccf6194-apiservice-cert\") pod \"metallb-operator-controller-manager-5d9448b85b-kszzj\" (UID: \"a4f19703-b458-4f44-b824-6c249ccf6194\") " pod="metallb-system/metallb-operator-controller-manager-5d9448b85b-kszzj" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.048789 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a4f19703-b458-4f44-b824-6c249ccf6194-webhook-cert\") pod \"metallb-operator-controller-manager-5d9448b85b-kszzj\" (UID: \"a4f19703-b458-4f44-b824-6c249ccf6194\") " pod="metallb-system/metallb-operator-controller-manager-5d9448b85b-kszzj" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.149916 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dfqp\" (UniqueName: \"kubernetes.io/projected/a4f19703-b458-4f44-b824-6c249ccf6194-kube-api-access-4dfqp\") pod \"metallb-operator-controller-manager-5d9448b85b-kszzj\" (UID: \"a4f19703-b458-4f44-b824-6c249ccf6194\") " pod="metallb-system/metallb-operator-controller-manager-5d9448b85b-kszzj" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.149969 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a4f19703-b458-4f44-b824-6c249ccf6194-apiservice-cert\") pod \"metallb-operator-controller-manager-5d9448b85b-kszzj\" (UID: \"a4f19703-b458-4f44-b824-6c249ccf6194\") " pod="metallb-system/metallb-operator-controller-manager-5d9448b85b-kszzj" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.149996 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a4f19703-b458-4f44-b824-6c249ccf6194-webhook-cert\") pod \"metallb-operator-controller-manager-5d9448b85b-kszzj\" (UID: \"a4f19703-b458-4f44-b824-6c249ccf6194\") " pod="metallb-system/metallb-operator-controller-manager-5d9448b85b-kszzj" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.159778 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a4f19703-b458-4f44-b824-6c249ccf6194-apiservice-cert\") pod \"metallb-operator-controller-manager-5d9448b85b-kszzj\" (UID: \"a4f19703-b458-4f44-b824-6c249ccf6194\") " pod="metallb-system/metallb-operator-controller-manager-5d9448b85b-kszzj" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.159794 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a4f19703-b458-4f44-b824-6c249ccf6194-webhook-cert\") pod \"metallb-operator-controller-manager-5d9448b85b-kszzj\" (UID: \"a4f19703-b458-4f44-b824-6c249ccf6194\") " pod="metallb-system/metallb-operator-controller-manager-5d9448b85b-kszzj" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.169672 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dfqp\" (UniqueName: \"kubernetes.io/projected/a4f19703-b458-4f44-b824-6c249ccf6194-kube-api-access-4dfqp\") pod \"metallb-operator-controller-manager-5d9448b85b-kszzj\" (UID: \"a4f19703-b458-4f44-b824-6c249ccf6194\") " pod="metallb-system/metallb-operator-controller-manager-5d9448b85b-kszzj" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.242988 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-f5b585465-glgcb"] Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.243673 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-f5b585465-glgcb" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.246359 4998 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-frspl" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.246497 4998 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.246536 4998 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.258743 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-f5b585465-glgcb"] Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.299521 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5d9448b85b-kszzj" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.352196 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/40a7ef48-e7b8-4668-8216-bb6392862991-apiservice-cert\") pod \"metallb-operator-webhook-server-f5b585465-glgcb\" (UID: \"40a7ef48-e7b8-4668-8216-bb6392862991\") " pod="metallb-system/metallb-operator-webhook-server-f5b585465-glgcb" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.352523 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/40a7ef48-e7b8-4668-8216-bb6392862991-webhook-cert\") pod \"metallb-operator-webhook-server-f5b585465-glgcb\" (UID: \"40a7ef48-e7b8-4668-8216-bb6392862991\") " pod="metallb-system/metallb-operator-webhook-server-f5b585465-glgcb" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.352784 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmcp7\" (UniqueName: \"kubernetes.io/projected/40a7ef48-e7b8-4668-8216-bb6392862991-kube-api-access-rmcp7\") pod \"metallb-operator-webhook-server-f5b585465-glgcb\" (UID: \"40a7ef48-e7b8-4668-8216-bb6392862991\") " pod="metallb-system/metallb-operator-webhook-server-f5b585465-glgcb" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.454715 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/40a7ef48-e7b8-4668-8216-bb6392862991-apiservice-cert\") pod \"metallb-operator-webhook-server-f5b585465-glgcb\" (UID: \"40a7ef48-e7b8-4668-8216-bb6392862991\") " pod="metallb-system/metallb-operator-webhook-server-f5b585465-glgcb" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.454805 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/40a7ef48-e7b8-4668-8216-bb6392862991-webhook-cert\") pod \"metallb-operator-webhook-server-f5b585465-glgcb\" (UID: \"40a7ef48-e7b8-4668-8216-bb6392862991\") " pod="metallb-system/metallb-operator-webhook-server-f5b585465-glgcb" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.454954 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmcp7\" (UniqueName: \"kubernetes.io/projected/40a7ef48-e7b8-4668-8216-bb6392862991-kube-api-access-rmcp7\") pod \"metallb-operator-webhook-server-f5b585465-glgcb\" (UID: \"40a7ef48-e7b8-4668-8216-bb6392862991\") " pod="metallb-system/metallb-operator-webhook-server-f5b585465-glgcb" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.467806 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/40a7ef48-e7b8-4668-8216-bb6392862991-webhook-cert\") pod \"metallb-operator-webhook-server-f5b585465-glgcb\" (UID: \"40a7ef48-e7b8-4668-8216-bb6392862991\") " pod="metallb-system/metallb-operator-webhook-server-f5b585465-glgcb" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.474172 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/40a7ef48-e7b8-4668-8216-bb6392862991-apiservice-cert\") pod \"metallb-operator-webhook-server-f5b585465-glgcb\" (UID: \"40a7ef48-e7b8-4668-8216-bb6392862991\") " pod="metallb-system/metallb-operator-webhook-server-f5b585465-glgcb" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.478480 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmcp7\" (UniqueName: \"kubernetes.io/projected/40a7ef48-e7b8-4668-8216-bb6392862991-kube-api-access-rmcp7\") pod \"metallb-operator-webhook-server-f5b585465-glgcb\" (UID: \"40a7ef48-e7b8-4668-8216-bb6392862991\") " pod="metallb-system/metallb-operator-webhook-server-f5b585465-glgcb" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.558788 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-f5b585465-glgcb" Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.701006 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5d9448b85b-kszzj"] Dec 03 16:17:58 crc kubenswrapper[4998]: W1203 16:17:58.713464 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4f19703_b458_4f44_b824_6c249ccf6194.slice/crio-77ea9251a90c0dd09d859d1dc00930def43eacad73517200a0d8a30755cdfee5 WatchSource:0}: Error finding container 77ea9251a90c0dd09d859d1dc00930def43eacad73517200a0d8a30755cdfee5: Status 404 returned error can't find the container with id 77ea9251a90c0dd09d859d1dc00930def43eacad73517200a0d8a30755cdfee5 Dec 03 16:17:58 crc kubenswrapper[4998]: I1203 16:17:58.990074 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-f5b585465-glgcb"] Dec 03 16:17:59 crc kubenswrapper[4998]: W1203 16:17:59.005664 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40a7ef48_e7b8_4668_8216_bb6392862991.slice/crio-ff693b490b760eb73d5b37a11a50b612140186222a58fd7076ac8678e1a1d1ca WatchSource:0}: Error finding container ff693b490b760eb73d5b37a11a50b612140186222a58fd7076ac8678e1a1d1ca: Status 404 returned error can't find the container with id ff693b490b760eb73d5b37a11a50b612140186222a58fd7076ac8678e1a1d1ca Dec 03 16:17:59 crc kubenswrapper[4998]: I1203 16:17:59.658917 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-f5b585465-glgcb" event={"ID":"40a7ef48-e7b8-4668-8216-bb6392862991","Type":"ContainerStarted","Data":"ff693b490b760eb73d5b37a11a50b612140186222a58fd7076ac8678e1a1d1ca"} Dec 03 16:17:59 crc kubenswrapper[4998]: I1203 16:17:59.661302 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5d9448b85b-kszzj" event={"ID":"a4f19703-b458-4f44-b824-6c249ccf6194","Type":"ContainerStarted","Data":"77ea9251a90c0dd09d859d1dc00930def43eacad73517200a0d8a30755cdfee5"} Dec 03 16:18:00 crc kubenswrapper[4998]: I1203 16:18:00.806217 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-z9mwr" Dec 03 16:18:00 crc kubenswrapper[4998]: I1203 16:18:00.806260 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-z9mwr" Dec 03 16:18:00 crc kubenswrapper[4998]: I1203 16:18:00.850381 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-z9mwr" Dec 03 16:18:01 crc kubenswrapper[4998]: I1203 16:18:01.719497 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-z9mwr" Dec 03 16:18:02 crc kubenswrapper[4998]: I1203 16:18:02.682927 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5d9448b85b-kszzj" event={"ID":"a4f19703-b458-4f44-b824-6c249ccf6194","Type":"ContainerStarted","Data":"9e1c92249490bcaf1b1b83395e7fac9635e6f7cc0e92514274c823916608167d"} Dec 03 16:18:02 crc kubenswrapper[4998]: I1203 16:18:02.683002 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5d9448b85b-kszzj" Dec 03 16:18:02 crc kubenswrapper[4998]: I1203 16:18:02.701942 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5d9448b85b-kszzj" podStartSLOduration=2.115365408 podStartE2EDuration="5.70192385s" podCreationTimestamp="2025-12-03 16:17:57 +0000 UTC" firstStartedPulling="2025-12-03 16:17:58.715580392 +0000 UTC m=+857.327280615" lastFinishedPulling="2025-12-03 16:18:02.302138834 +0000 UTC m=+860.913839057" observedRunningTime="2025-12-03 16:18:02.698619267 +0000 UTC m=+861.310319490" watchObservedRunningTime="2025-12-03 16:18:02.70192385 +0000 UTC m=+861.313624073" Dec 03 16:18:02 crc kubenswrapper[4998]: I1203 16:18:02.874717 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z9mwr"] Dec 03 16:18:03 crc kubenswrapper[4998]: I1203 16:18:03.685838 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-z9mwr" podUID="e5c42d90-b0b9-4ac4-a5d9-d355a5d27388" containerName="registry-server" containerID="cri-o://c090b865321578dcd456e89110d9af6321784b65249827f432217c12a183fd4d" gracePeriod=2 Dec 03 16:18:04 crc kubenswrapper[4998]: I1203 16:18:04.696483 4998 generic.go:334] "Generic (PLEG): container finished" podID="e5c42d90-b0b9-4ac4-a5d9-d355a5d27388" containerID="c090b865321578dcd456e89110d9af6321784b65249827f432217c12a183fd4d" exitCode=0 Dec 03 16:18:04 crc kubenswrapper[4998]: I1203 16:18:04.696543 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z9mwr" event={"ID":"e5c42d90-b0b9-4ac4-a5d9-d355a5d27388","Type":"ContainerDied","Data":"c090b865321578dcd456e89110d9af6321784b65249827f432217c12a183fd4d"} Dec 03 16:18:05 crc kubenswrapper[4998]: I1203 16:18:05.385132 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z9mwr" Dec 03 16:18:05 crc kubenswrapper[4998]: I1203 16:18:05.453064 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5c42d90-b0b9-4ac4-a5d9-d355a5d27388-catalog-content\") pod \"e5c42d90-b0b9-4ac4-a5d9-d355a5d27388\" (UID: \"e5c42d90-b0b9-4ac4-a5d9-d355a5d27388\") " Dec 03 16:18:05 crc kubenswrapper[4998]: I1203 16:18:05.453447 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5c42d90-b0b9-4ac4-a5d9-d355a5d27388-utilities\") pod \"e5c42d90-b0b9-4ac4-a5d9-d355a5d27388\" (UID: \"e5c42d90-b0b9-4ac4-a5d9-d355a5d27388\") " Dec 03 16:18:05 crc kubenswrapper[4998]: I1203 16:18:05.453717 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88rds\" (UniqueName: \"kubernetes.io/projected/e5c42d90-b0b9-4ac4-a5d9-d355a5d27388-kube-api-access-88rds\") pod \"e5c42d90-b0b9-4ac4-a5d9-d355a5d27388\" (UID: \"e5c42d90-b0b9-4ac4-a5d9-d355a5d27388\") " Dec 03 16:18:05 crc kubenswrapper[4998]: I1203 16:18:05.454888 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5c42d90-b0b9-4ac4-a5d9-d355a5d27388-utilities" (OuterVolumeSpecName: "utilities") pod "e5c42d90-b0b9-4ac4-a5d9-d355a5d27388" (UID: "e5c42d90-b0b9-4ac4-a5d9-d355a5d27388"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:18:05 crc kubenswrapper[4998]: I1203 16:18:05.460717 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5c42d90-b0b9-4ac4-a5d9-d355a5d27388-kube-api-access-88rds" (OuterVolumeSpecName: "kube-api-access-88rds") pod "e5c42d90-b0b9-4ac4-a5d9-d355a5d27388" (UID: "e5c42d90-b0b9-4ac4-a5d9-d355a5d27388"). InnerVolumeSpecName "kube-api-access-88rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:18:05 crc kubenswrapper[4998]: I1203 16:18:05.506009 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5c42d90-b0b9-4ac4-a5d9-d355a5d27388-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5c42d90-b0b9-4ac4-a5d9-d355a5d27388" (UID: "e5c42d90-b0b9-4ac4-a5d9-d355a5d27388"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:18:05 crc kubenswrapper[4998]: I1203 16:18:05.555536 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5c42d90-b0b9-4ac4-a5d9-d355a5d27388-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:18:05 crc kubenswrapper[4998]: I1203 16:18:05.555569 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88rds\" (UniqueName: \"kubernetes.io/projected/e5c42d90-b0b9-4ac4-a5d9-d355a5d27388-kube-api-access-88rds\") on node \"crc\" DevicePath \"\"" Dec 03 16:18:05 crc kubenswrapper[4998]: I1203 16:18:05.555578 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5c42d90-b0b9-4ac4-a5d9-d355a5d27388-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:18:05 crc kubenswrapper[4998]: I1203 16:18:05.706517 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-f5b585465-glgcb" event={"ID":"40a7ef48-e7b8-4668-8216-bb6392862991","Type":"ContainerStarted","Data":"7863f287690c7eddce84ff1bdd216a01fe7a230dea5933014250d95833947274"} Dec 03 16:18:05 crc kubenswrapper[4998]: I1203 16:18:05.706693 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-f5b585465-glgcb" Dec 03 16:18:05 crc kubenswrapper[4998]: I1203 16:18:05.709528 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z9mwr" event={"ID":"e5c42d90-b0b9-4ac4-a5d9-d355a5d27388","Type":"ContainerDied","Data":"bb8aa35fd610e5ab7d9a397f9fe5eec674830e08367d94708834c2f6fae39734"} Dec 03 16:18:05 crc kubenswrapper[4998]: I1203 16:18:05.709584 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z9mwr" Dec 03 16:18:05 crc kubenswrapper[4998]: I1203 16:18:05.709595 4998 scope.go:117] "RemoveContainer" containerID="c090b865321578dcd456e89110d9af6321784b65249827f432217c12a183fd4d" Dec 03 16:18:05 crc kubenswrapper[4998]: I1203 16:18:05.726700 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-f5b585465-glgcb" podStartSLOduration=1.322731226 podStartE2EDuration="7.726668073s" podCreationTimestamp="2025-12-03 16:17:58 +0000 UTC" firstStartedPulling="2025-12-03 16:17:59.008228217 +0000 UTC m=+857.619928450" lastFinishedPulling="2025-12-03 16:18:05.412165074 +0000 UTC m=+864.023865297" observedRunningTime="2025-12-03 16:18:05.72112867 +0000 UTC m=+864.332828903" watchObservedRunningTime="2025-12-03 16:18:05.726668073 +0000 UTC m=+864.338368286" Dec 03 16:18:05 crc kubenswrapper[4998]: I1203 16:18:05.737525 4998 scope.go:117] "RemoveContainer" containerID="59febff8a36410df271852bdbf7fee5eef27e88a0711f2ce433a3bc1183aa7d8" Dec 03 16:18:05 crc kubenswrapper[4998]: I1203 16:18:05.746055 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z9mwr"] Dec 03 16:18:05 crc kubenswrapper[4998]: I1203 16:18:05.752848 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-z9mwr"] Dec 03 16:18:05 crc kubenswrapper[4998]: I1203 16:18:05.762036 4998 scope.go:117] "RemoveContainer" containerID="09188b4a9d7e58c5a70bbefbf7122a0c00dd239bf19a8eece428f338763fa8f6" Dec 03 16:18:07 crc kubenswrapper[4998]: I1203 16:18:07.687008 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5c42d90-b0b9-4ac4-a5d9-d355a5d27388" path="/var/lib/kubelet/pods/e5c42d90-b0b9-4ac4-a5d9-d355a5d27388/volumes" Dec 03 16:18:18 crc kubenswrapper[4998]: I1203 16:18:18.565699 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-f5b585465-glgcb" Dec 03 16:18:27 crc kubenswrapper[4998]: I1203 16:18:27.111292 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:18:27 crc kubenswrapper[4998]: I1203 16:18:27.112038 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:18:38 crc kubenswrapper[4998]: I1203 16:18:38.303531 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5d9448b85b-kszzj" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.054368 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-m449d"] Dec 03 16:18:39 crc kubenswrapper[4998]: E1203 16:18:39.054626 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5c42d90-b0b9-4ac4-a5d9-d355a5d27388" containerName="registry-server" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.054642 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5c42d90-b0b9-4ac4-a5d9-d355a5d27388" containerName="registry-server" Dec 03 16:18:39 crc kubenswrapper[4998]: E1203 16:18:39.054658 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5c42d90-b0b9-4ac4-a5d9-d355a5d27388" containerName="extract-utilities" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.054667 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5c42d90-b0b9-4ac4-a5d9-d355a5d27388" containerName="extract-utilities" Dec 03 16:18:39 crc kubenswrapper[4998]: E1203 16:18:39.054683 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5c42d90-b0b9-4ac4-a5d9-d355a5d27388" containerName="extract-content" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.054691 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5c42d90-b0b9-4ac4-a5d9-d355a5d27388" containerName="extract-content" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.054845 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5c42d90-b0b9-4ac4-a5d9-d355a5d27388" containerName="registry-server" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.057682 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.060159 4998 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.061881 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.062108 4998 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-d5mrt" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.086293 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-j8sl8"] Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.087968 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j8sl8" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.092178 4998 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.096961 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-j8sl8"] Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.104682 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rpwl\" (UniqueName: \"kubernetes.io/projected/45a612f3-9622-4ee1-be7e-04e7c5035f7a-kube-api-access-7rpwl\") pod \"frr-k8s-m449d\" (UID: \"45a612f3-9622-4ee1-be7e-04e7c5035f7a\") " pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.104727 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/45a612f3-9622-4ee1-be7e-04e7c5035f7a-frr-conf\") pod \"frr-k8s-m449d\" (UID: \"45a612f3-9622-4ee1-be7e-04e7c5035f7a\") " pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.104750 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/45a612f3-9622-4ee1-be7e-04e7c5035f7a-frr-sockets\") pod \"frr-k8s-m449d\" (UID: \"45a612f3-9622-4ee1-be7e-04e7c5035f7a\") " pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.104908 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/45a612f3-9622-4ee1-be7e-04e7c5035f7a-metrics\") pod \"frr-k8s-m449d\" (UID: \"45a612f3-9622-4ee1-be7e-04e7c5035f7a\") " pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.105019 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/45a612f3-9622-4ee1-be7e-04e7c5035f7a-frr-startup\") pod \"frr-k8s-m449d\" (UID: \"45a612f3-9622-4ee1-be7e-04e7c5035f7a\") " pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.105048 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/45a612f3-9622-4ee1-be7e-04e7c5035f7a-reloader\") pod \"frr-k8s-m449d\" (UID: \"45a612f3-9622-4ee1-be7e-04e7c5035f7a\") " pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.105095 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/45a612f3-9622-4ee1-be7e-04e7c5035f7a-metrics-certs\") pod \"frr-k8s-m449d\" (UID: \"45a612f3-9622-4ee1-be7e-04e7c5035f7a\") " pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.179165 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-qd4fq"] Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.180218 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qd4fq" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.185590 4998 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.185653 4998 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.185704 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.185711 4998 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-nbv6d" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.206606 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/45a612f3-9622-4ee1-be7e-04e7c5035f7a-metrics-certs\") pod \"frr-k8s-m449d\" (UID: \"45a612f3-9622-4ee1-be7e-04e7c5035f7a\") " pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.206877 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rpwl\" (UniqueName: \"kubernetes.io/projected/45a612f3-9622-4ee1-be7e-04e7c5035f7a-kube-api-access-7rpwl\") pod \"frr-k8s-m449d\" (UID: \"45a612f3-9622-4ee1-be7e-04e7c5035f7a\") " pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.207007 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/45a612f3-9622-4ee1-be7e-04e7c5035f7a-frr-conf\") pod \"frr-k8s-m449d\" (UID: \"45a612f3-9622-4ee1-be7e-04e7c5035f7a\") " pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.207435 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/45a612f3-9622-4ee1-be7e-04e7c5035f7a-frr-sockets\") pod \"frr-k8s-m449d\" (UID: \"45a612f3-9622-4ee1-be7e-04e7c5035f7a\") " pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.207941 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmvx9\" (UniqueName: \"kubernetes.io/projected/3830057a-341e-4586-8f1a-f80d34f8c997-kube-api-access-tmvx9\") pod \"speaker-qd4fq\" (UID: \"3830057a-341e-4586-8f1a-f80d34f8c997\") " pod="metallb-system/speaker-qd4fq" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.208071 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cdd57e8f-0d60-4634-8460-2bee971c0bc7-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-j8sl8\" (UID: \"cdd57e8f-0d60-4634-8460-2bee971c0bc7\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j8sl8" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.208228 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/45a612f3-9622-4ee1-be7e-04e7c5035f7a-metrics\") pod \"frr-k8s-m449d\" (UID: \"45a612f3-9622-4ee1-be7e-04e7c5035f7a\") " pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.208532 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/45a612f3-9622-4ee1-be7e-04e7c5035f7a-frr-startup\") pod \"frr-k8s-m449d\" (UID: \"45a612f3-9622-4ee1-be7e-04e7c5035f7a\") " pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.208605 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/45a612f3-9622-4ee1-be7e-04e7c5035f7a-reloader\") pod \"frr-k8s-m449d\" (UID: \"45a612f3-9622-4ee1-be7e-04e7c5035f7a\") " pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.208680 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/3830057a-341e-4586-8f1a-f80d34f8c997-memberlist\") pod \"speaker-qd4fq\" (UID: \"3830057a-341e-4586-8f1a-f80d34f8c997\") " pod="metallb-system/speaker-qd4fq" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.208769 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/3830057a-341e-4586-8f1a-f80d34f8c997-metallb-excludel2\") pod \"speaker-qd4fq\" (UID: \"3830057a-341e-4586-8f1a-f80d34f8c997\") " pod="metallb-system/speaker-qd4fq" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.208845 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zf2bk\" (UniqueName: \"kubernetes.io/projected/cdd57e8f-0d60-4634-8460-2bee971c0bc7-kube-api-access-zf2bk\") pod \"frr-k8s-webhook-server-7fcb986d4-j8sl8\" (UID: \"cdd57e8f-0d60-4634-8460-2bee971c0bc7\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j8sl8" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.209254 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3830057a-341e-4586-8f1a-f80d34f8c997-metrics-certs\") pod \"speaker-qd4fq\" (UID: \"3830057a-341e-4586-8f1a-f80d34f8c997\") " pod="metallb-system/speaker-qd4fq" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.207894 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/45a612f3-9622-4ee1-be7e-04e7c5035f7a-frr-sockets\") pod \"frr-k8s-m449d\" (UID: \"45a612f3-9622-4ee1-be7e-04e7c5035f7a\") " pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.208420 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/45a612f3-9622-4ee1-be7e-04e7c5035f7a-metrics\") pod \"frr-k8s-m449d\" (UID: \"45a612f3-9622-4ee1-be7e-04e7c5035f7a\") " pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.207387 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/45a612f3-9622-4ee1-be7e-04e7c5035f7a-frr-conf\") pod \"frr-k8s-m449d\" (UID: \"45a612f3-9622-4ee1-be7e-04e7c5035f7a\") " pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.210180 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/45a612f3-9622-4ee1-be7e-04e7c5035f7a-reloader\") pod \"frr-k8s-m449d\" (UID: \"45a612f3-9622-4ee1-be7e-04e7c5035f7a\") " pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.210322 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-d9v5x"] Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.211472 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-d9v5x" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.212746 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/45a612f3-9622-4ee1-be7e-04e7c5035f7a-frr-startup\") pod \"frr-k8s-m449d\" (UID: \"45a612f3-9622-4ee1-be7e-04e7c5035f7a\") " pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.214305 4998 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.225260 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/45a612f3-9622-4ee1-be7e-04e7c5035f7a-metrics-certs\") pod \"frr-k8s-m449d\" (UID: \"45a612f3-9622-4ee1-be7e-04e7c5035f7a\") " pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.228150 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rpwl\" (UniqueName: \"kubernetes.io/projected/45a612f3-9622-4ee1-be7e-04e7c5035f7a-kube-api-access-7rpwl\") pod \"frr-k8s-m449d\" (UID: \"45a612f3-9622-4ee1-be7e-04e7c5035f7a\") " pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.244811 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-d9v5x"] Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.310669 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9ce7261b-c84b-473e-a501-b570a4ba5b02-cert\") pod \"controller-f8648f98b-d9v5x\" (UID: \"9ce7261b-c84b-473e-a501-b570a4ba5b02\") " pod="metallb-system/controller-f8648f98b-d9v5x" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.310834 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l52v\" (UniqueName: \"kubernetes.io/projected/9ce7261b-c84b-473e-a501-b570a4ba5b02-kube-api-access-2l52v\") pod \"controller-f8648f98b-d9v5x\" (UID: \"9ce7261b-c84b-473e-a501-b570a4ba5b02\") " pod="metallb-system/controller-f8648f98b-d9v5x" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.310878 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zf2bk\" (UniqueName: \"kubernetes.io/projected/cdd57e8f-0d60-4634-8460-2bee971c0bc7-kube-api-access-zf2bk\") pod \"frr-k8s-webhook-server-7fcb986d4-j8sl8\" (UID: \"cdd57e8f-0d60-4634-8460-2bee971c0bc7\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j8sl8" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.310898 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/3830057a-341e-4586-8f1a-f80d34f8c997-memberlist\") pod \"speaker-qd4fq\" (UID: \"3830057a-341e-4586-8f1a-f80d34f8c997\") " pod="metallb-system/speaker-qd4fq" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.310913 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/3830057a-341e-4586-8f1a-f80d34f8c997-metallb-excludel2\") pod \"speaker-qd4fq\" (UID: \"3830057a-341e-4586-8f1a-f80d34f8c997\") " pod="metallb-system/speaker-qd4fq" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.310929 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ce7261b-c84b-473e-a501-b570a4ba5b02-metrics-certs\") pod \"controller-f8648f98b-d9v5x\" (UID: \"9ce7261b-c84b-473e-a501-b570a4ba5b02\") " pod="metallb-system/controller-f8648f98b-d9v5x" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.310947 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3830057a-341e-4586-8f1a-f80d34f8c997-metrics-certs\") pod \"speaker-qd4fq\" (UID: \"3830057a-341e-4586-8f1a-f80d34f8c997\") " pod="metallb-system/speaker-qd4fq" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.310991 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmvx9\" (UniqueName: \"kubernetes.io/projected/3830057a-341e-4586-8f1a-f80d34f8c997-kube-api-access-tmvx9\") pod \"speaker-qd4fq\" (UID: \"3830057a-341e-4586-8f1a-f80d34f8c997\") " pod="metallb-system/speaker-qd4fq" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.311013 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cdd57e8f-0d60-4634-8460-2bee971c0bc7-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-j8sl8\" (UID: \"cdd57e8f-0d60-4634-8460-2bee971c0bc7\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j8sl8" Dec 03 16:18:39 crc kubenswrapper[4998]: E1203 16:18:39.311064 4998 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 16:18:39 crc kubenswrapper[4998]: E1203 16:18:39.311117 4998 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 03 16:18:39 crc kubenswrapper[4998]: E1203 16:18:39.311144 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3830057a-341e-4586-8f1a-f80d34f8c997-memberlist podName:3830057a-341e-4586-8f1a-f80d34f8c997 nodeName:}" failed. No retries permitted until 2025-12-03 16:18:39.811125645 +0000 UTC m=+898.422825858 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/3830057a-341e-4586-8f1a-f80d34f8c997-memberlist") pod "speaker-qd4fq" (UID: "3830057a-341e-4586-8f1a-f80d34f8c997") : secret "metallb-memberlist" not found Dec 03 16:18:39 crc kubenswrapper[4998]: E1203 16:18:39.311177 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3830057a-341e-4586-8f1a-f80d34f8c997-metrics-certs podName:3830057a-341e-4586-8f1a-f80d34f8c997 nodeName:}" failed. No retries permitted until 2025-12-03 16:18:39.811160485 +0000 UTC m=+898.422860708 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3830057a-341e-4586-8f1a-f80d34f8c997-metrics-certs") pod "speaker-qd4fq" (UID: "3830057a-341e-4586-8f1a-f80d34f8c997") : secret "speaker-certs-secret" not found Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.311796 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/3830057a-341e-4586-8f1a-f80d34f8c997-metallb-excludel2\") pod \"speaker-qd4fq\" (UID: \"3830057a-341e-4586-8f1a-f80d34f8c997\") " pod="metallb-system/speaker-qd4fq" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.313865 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cdd57e8f-0d60-4634-8460-2bee971c0bc7-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-j8sl8\" (UID: \"cdd57e8f-0d60-4634-8460-2bee971c0bc7\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j8sl8" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.326159 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmvx9\" (UniqueName: \"kubernetes.io/projected/3830057a-341e-4586-8f1a-f80d34f8c997-kube-api-access-tmvx9\") pod \"speaker-qd4fq\" (UID: \"3830057a-341e-4586-8f1a-f80d34f8c997\") " pod="metallb-system/speaker-qd4fq" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.327444 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zf2bk\" (UniqueName: \"kubernetes.io/projected/cdd57e8f-0d60-4634-8460-2bee971c0bc7-kube-api-access-zf2bk\") pod \"frr-k8s-webhook-server-7fcb986d4-j8sl8\" (UID: \"cdd57e8f-0d60-4634-8460-2bee971c0bc7\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j8sl8" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.388588 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.405622 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j8sl8" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.413365 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l52v\" (UniqueName: \"kubernetes.io/projected/9ce7261b-c84b-473e-a501-b570a4ba5b02-kube-api-access-2l52v\") pod \"controller-f8648f98b-d9v5x\" (UID: \"9ce7261b-c84b-473e-a501-b570a4ba5b02\") " pod="metallb-system/controller-f8648f98b-d9v5x" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.413427 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ce7261b-c84b-473e-a501-b570a4ba5b02-metrics-certs\") pod \"controller-f8648f98b-d9v5x\" (UID: \"9ce7261b-c84b-473e-a501-b570a4ba5b02\") " pod="metallb-system/controller-f8648f98b-d9v5x" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.413489 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9ce7261b-c84b-473e-a501-b570a4ba5b02-cert\") pod \"controller-f8648f98b-d9v5x\" (UID: \"9ce7261b-c84b-473e-a501-b570a4ba5b02\") " pod="metallb-system/controller-f8648f98b-d9v5x" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.416608 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9ce7261b-c84b-473e-a501-b570a4ba5b02-cert\") pod \"controller-f8648f98b-d9v5x\" (UID: \"9ce7261b-c84b-473e-a501-b570a4ba5b02\") " pod="metallb-system/controller-f8648f98b-d9v5x" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.421264 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ce7261b-c84b-473e-a501-b570a4ba5b02-metrics-certs\") pod \"controller-f8648f98b-d9v5x\" (UID: \"9ce7261b-c84b-473e-a501-b570a4ba5b02\") " pod="metallb-system/controller-f8648f98b-d9v5x" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.430460 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l52v\" (UniqueName: \"kubernetes.io/projected/9ce7261b-c84b-473e-a501-b570a4ba5b02-kube-api-access-2l52v\") pod \"controller-f8648f98b-d9v5x\" (UID: \"9ce7261b-c84b-473e-a501-b570a4ba5b02\") " pod="metallb-system/controller-f8648f98b-d9v5x" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.579361 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-d9v5x" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.670091 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-j8sl8"] Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.829159 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/3830057a-341e-4586-8f1a-f80d34f8c997-memberlist\") pod \"speaker-qd4fq\" (UID: \"3830057a-341e-4586-8f1a-f80d34f8c997\") " pod="metallb-system/speaker-qd4fq" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.829546 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3830057a-341e-4586-8f1a-f80d34f8c997-metrics-certs\") pod \"speaker-qd4fq\" (UID: \"3830057a-341e-4586-8f1a-f80d34f8c997\") " pod="metallb-system/speaker-qd4fq" Dec 03 16:18:39 crc kubenswrapper[4998]: E1203 16:18:39.829354 4998 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 16:18:39 crc kubenswrapper[4998]: E1203 16:18:39.829736 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3830057a-341e-4586-8f1a-f80d34f8c997-memberlist podName:3830057a-341e-4586-8f1a-f80d34f8c997 nodeName:}" failed. No retries permitted until 2025-12-03 16:18:40.829707778 +0000 UTC m=+899.441408041 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/3830057a-341e-4586-8f1a-f80d34f8c997-memberlist") pod "speaker-qd4fq" (UID: "3830057a-341e-4586-8f1a-f80d34f8c997") : secret "metallb-memberlist" not found Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.838534 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3830057a-341e-4586-8f1a-f80d34f8c997-metrics-certs\") pod \"speaker-qd4fq\" (UID: \"3830057a-341e-4586-8f1a-f80d34f8c997\") " pod="metallb-system/speaker-qd4fq" Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.999059 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-d9v5x"] Dec 03 16:18:39 crc kubenswrapper[4998]: I1203 16:18:39.999473 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j8sl8" event={"ID":"cdd57e8f-0d60-4634-8460-2bee971c0bc7","Type":"ContainerStarted","Data":"0c1e6c1decc126535cdfa95beef143ed0b0bc69bf62be1edd282b9e5f8a28176"} Dec 03 16:18:40 crc kubenswrapper[4998]: I1203 16:18:40.002081 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-m449d" event={"ID":"45a612f3-9622-4ee1-be7e-04e7c5035f7a","Type":"ContainerStarted","Data":"f6478224911a925ce90afcbc2582d709377c01da655245eee4a337e68011990c"} Dec 03 16:18:40 crc kubenswrapper[4998]: W1203 16:18:40.003285 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ce7261b_c84b_473e_a501_b570a4ba5b02.slice/crio-ca9224e9e81d2647d24cdcc0731dc581e07e3107f25f507e42b6525ea7e5b554 WatchSource:0}: Error finding container ca9224e9e81d2647d24cdcc0731dc581e07e3107f25f507e42b6525ea7e5b554: Status 404 returned error can't find the container with id ca9224e9e81d2647d24cdcc0731dc581e07e3107f25f507e42b6525ea7e5b554 Dec 03 16:18:40 crc kubenswrapper[4998]: I1203 16:18:40.847346 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/3830057a-341e-4586-8f1a-f80d34f8c997-memberlist\") pod \"speaker-qd4fq\" (UID: \"3830057a-341e-4586-8f1a-f80d34f8c997\") " pod="metallb-system/speaker-qd4fq" Dec 03 16:18:40 crc kubenswrapper[4998]: I1203 16:18:40.852351 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/3830057a-341e-4586-8f1a-f80d34f8c997-memberlist\") pod \"speaker-qd4fq\" (UID: \"3830057a-341e-4586-8f1a-f80d34f8c997\") " pod="metallb-system/speaker-qd4fq" Dec 03 16:18:40 crc kubenswrapper[4998]: I1203 16:18:40.999040 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qd4fq" Dec 03 16:18:41 crc kubenswrapper[4998]: I1203 16:18:41.010279 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-d9v5x" event={"ID":"9ce7261b-c84b-473e-a501-b570a4ba5b02","Type":"ContainerStarted","Data":"176463afa2781e63bb46854a15c6fcb193dc9b9a94dc293f16b67b92ed9daf8b"} Dec 03 16:18:41 crc kubenswrapper[4998]: I1203 16:18:41.010322 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-d9v5x" event={"ID":"9ce7261b-c84b-473e-a501-b570a4ba5b02","Type":"ContainerStarted","Data":"a0806e1d87bdad75e52eff9bbec7697dcff389f7fa11ff2286688d68860c760e"} Dec 03 16:18:41 crc kubenswrapper[4998]: I1203 16:18:41.010333 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-d9v5x" event={"ID":"9ce7261b-c84b-473e-a501-b570a4ba5b02","Type":"ContainerStarted","Data":"ca9224e9e81d2647d24cdcc0731dc581e07e3107f25f507e42b6525ea7e5b554"} Dec 03 16:18:41 crc kubenswrapper[4998]: I1203 16:18:41.010453 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-d9v5x" Dec 03 16:18:41 crc kubenswrapper[4998]: W1203 16:18:41.038563 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3830057a_341e_4586_8f1a_f80d34f8c997.slice/crio-5193e83e6616c04171b27192ed032ede0ad3a8a78d069791ceffed8f7644e364 WatchSource:0}: Error finding container 5193e83e6616c04171b27192ed032ede0ad3a8a78d069791ceffed8f7644e364: Status 404 returned error can't find the container with id 5193e83e6616c04171b27192ed032ede0ad3a8a78d069791ceffed8f7644e364 Dec 03 16:18:41 crc kubenswrapper[4998]: I1203 16:18:41.042362 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-d9v5x" podStartSLOduration=2.042343948 podStartE2EDuration="2.042343948s" podCreationTimestamp="2025-12-03 16:18:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:18:41.03387871 +0000 UTC m=+899.645578973" watchObservedRunningTime="2025-12-03 16:18:41.042343948 +0000 UTC m=+899.654044171" Dec 03 16:18:42 crc kubenswrapper[4998]: I1203 16:18:42.028418 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qd4fq" event={"ID":"3830057a-341e-4586-8f1a-f80d34f8c997","Type":"ContainerStarted","Data":"d27cf9535f4d333f421ee9774b867051939db97cf76175c379409ad6fcb56010"} Dec 03 16:18:42 crc kubenswrapper[4998]: I1203 16:18:42.028813 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qd4fq" event={"ID":"3830057a-341e-4586-8f1a-f80d34f8c997","Type":"ContainerStarted","Data":"8fd61eec2dea8d0b8f29e491a2fdf291b9ca0000686f575e48b6cf11ddae0a39"} Dec 03 16:18:42 crc kubenswrapper[4998]: I1203 16:18:42.028830 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qd4fq" event={"ID":"3830057a-341e-4586-8f1a-f80d34f8c997","Type":"ContainerStarted","Data":"5193e83e6616c04171b27192ed032ede0ad3a8a78d069791ceffed8f7644e364"} Dec 03 16:18:42 crc kubenswrapper[4998]: I1203 16:18:42.029014 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-qd4fq" Dec 03 16:18:42 crc kubenswrapper[4998]: I1203 16:18:42.051080 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-qd4fq" podStartSLOduration=3.051065816 podStartE2EDuration="3.051065816s" podCreationTimestamp="2025-12-03 16:18:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:18:42.048663183 +0000 UTC m=+900.660363406" watchObservedRunningTime="2025-12-03 16:18:42.051065816 +0000 UTC m=+900.662766039" Dec 03 16:18:47 crc kubenswrapper[4998]: I1203 16:18:47.072225 4998 generic.go:334] "Generic (PLEG): container finished" podID="45a612f3-9622-4ee1-be7e-04e7c5035f7a" containerID="5aa4ab31a81d0a9d8f4770ce7336e13142d4328e2ee0921588938c1c596e020c" exitCode=0 Dec 03 16:18:47 crc kubenswrapper[4998]: I1203 16:18:47.072291 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-m449d" event={"ID":"45a612f3-9622-4ee1-be7e-04e7c5035f7a","Type":"ContainerDied","Data":"5aa4ab31a81d0a9d8f4770ce7336e13142d4328e2ee0921588938c1c596e020c"} Dec 03 16:18:47 crc kubenswrapper[4998]: I1203 16:18:47.076319 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j8sl8" event={"ID":"cdd57e8f-0d60-4634-8460-2bee971c0bc7","Type":"ContainerStarted","Data":"ccb2e5563ccd7799c19dd5cd8b97cde005a3ab7c292db6629811f332680663a7"} Dec 03 16:18:47 crc kubenswrapper[4998]: I1203 16:18:47.076503 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j8sl8" Dec 03 16:18:48 crc kubenswrapper[4998]: I1203 16:18:48.085806 4998 generic.go:334] "Generic (PLEG): container finished" podID="45a612f3-9622-4ee1-be7e-04e7c5035f7a" containerID="94507b9e2ae0a081d2a82acdee17f7d5438f2be4f0f0c88401c6ffc852d978a4" exitCode=0 Dec 03 16:18:48 crc kubenswrapper[4998]: I1203 16:18:48.085891 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-m449d" event={"ID":"45a612f3-9622-4ee1-be7e-04e7c5035f7a","Type":"ContainerDied","Data":"94507b9e2ae0a081d2a82acdee17f7d5438f2be4f0f0c88401c6ffc852d978a4"} Dec 03 16:18:48 crc kubenswrapper[4998]: I1203 16:18:48.121522 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j8sl8" podStartSLOduration=2.265841483 podStartE2EDuration="9.121489444s" podCreationTimestamp="2025-12-03 16:18:39 +0000 UTC" firstStartedPulling="2025-12-03 16:18:39.685528809 +0000 UTC m=+898.297229052" lastFinishedPulling="2025-12-03 16:18:46.54117679 +0000 UTC m=+905.152877013" observedRunningTime="2025-12-03 16:18:47.127165296 +0000 UTC m=+905.738865519" watchObservedRunningTime="2025-12-03 16:18:48.121489444 +0000 UTC m=+906.733189707" Dec 03 16:18:49 crc kubenswrapper[4998]: I1203 16:18:49.094450 4998 generic.go:334] "Generic (PLEG): container finished" podID="45a612f3-9622-4ee1-be7e-04e7c5035f7a" containerID="7578b4604d65ad8c72642b83ca48dea4b4038c81f2426effcb4b72d99ad6ac69" exitCode=0 Dec 03 16:18:49 crc kubenswrapper[4998]: I1203 16:18:49.094505 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-m449d" event={"ID":"45a612f3-9622-4ee1-be7e-04e7c5035f7a","Type":"ContainerDied","Data":"7578b4604d65ad8c72642b83ca48dea4b4038c81f2426effcb4b72d99ad6ac69"} Dec 03 16:18:50 crc kubenswrapper[4998]: I1203 16:18:50.105506 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-m449d" event={"ID":"45a612f3-9622-4ee1-be7e-04e7c5035f7a","Type":"ContainerStarted","Data":"52767b8d09588970d48c2d566812cb4ec6c531a64998347870152c55f720ca63"} Dec 03 16:18:50 crc kubenswrapper[4998]: I1203 16:18:50.106549 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-m449d" event={"ID":"45a612f3-9622-4ee1-be7e-04e7c5035f7a","Type":"ContainerStarted","Data":"fa18e281fad3749c8df0cb01bae5692db6ac4b2fd9028aad1dd4155feba92f52"} Dec 03 16:18:51 crc kubenswrapper[4998]: I1203 16:18:51.131237 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-m449d" event={"ID":"45a612f3-9622-4ee1-be7e-04e7c5035f7a","Type":"ContainerStarted","Data":"7d9b9f212a6c67881b6ae9ec3b0861e3742f03c1874bf19fdcdc28e560726498"} Dec 03 16:18:51 crc kubenswrapper[4998]: I1203 16:18:51.131275 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-m449d" event={"ID":"45a612f3-9622-4ee1-be7e-04e7c5035f7a","Type":"ContainerStarted","Data":"64c36e48b46ab0038e782251a0f1f8e72e04389b00fc56c464fc4c885c3eeb6f"} Dec 03 16:18:51 crc kubenswrapper[4998]: I1203 16:18:51.131285 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-m449d" event={"ID":"45a612f3-9622-4ee1-be7e-04e7c5035f7a","Type":"ContainerStarted","Data":"a0784dcdc64b7fabe0b514dfd684ce0f8d60a675ae77335b8d9ca517a8d47839"} Dec 03 16:18:52 crc kubenswrapper[4998]: I1203 16:18:52.144560 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-m449d" event={"ID":"45a612f3-9622-4ee1-be7e-04e7c5035f7a","Type":"ContainerStarted","Data":"8c4ed26763d7358e89614e094808917c98e00fb1306922aa0e58a45d41136869"} Dec 03 16:18:52 crc kubenswrapper[4998]: I1203 16:18:52.145018 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:52 crc kubenswrapper[4998]: I1203 16:18:52.180527 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-m449d" podStartSLOduration=6.283177865 podStartE2EDuration="13.1805074s" podCreationTimestamp="2025-12-03 16:18:39 +0000 UTC" firstStartedPulling="2025-12-03 16:18:39.623917685 +0000 UTC m=+898.235617908" lastFinishedPulling="2025-12-03 16:18:46.52124718 +0000 UTC m=+905.132947443" observedRunningTime="2025-12-03 16:18:52.171942071 +0000 UTC m=+910.783642324" watchObservedRunningTime="2025-12-03 16:18:52.1805074 +0000 UTC m=+910.792207633" Dec 03 16:18:54 crc kubenswrapper[4998]: I1203 16:18:54.389580 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:54 crc kubenswrapper[4998]: I1203 16:18:54.438165 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:57 crc kubenswrapper[4998]: I1203 16:18:57.111578 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:18:57 crc kubenswrapper[4998]: I1203 16:18:57.111710 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:18:59 crc kubenswrapper[4998]: I1203 16:18:59.391196 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-m449d" Dec 03 16:18:59 crc kubenswrapper[4998]: I1203 16:18:59.414078 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-j8sl8" Dec 03 16:18:59 crc kubenswrapper[4998]: I1203 16:18:59.587589 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-d9v5x" Dec 03 16:19:01 crc kubenswrapper[4998]: I1203 16:19:01.003127 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-qd4fq" Dec 03 16:19:04 crc kubenswrapper[4998]: I1203 16:19:04.204976 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-zdxk7"] Dec 03 16:19:04 crc kubenswrapper[4998]: I1203 16:19:04.206243 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zdxk7" Dec 03 16:19:04 crc kubenswrapper[4998]: I1203 16:19:04.209098 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-gpgfs" Dec 03 16:19:04 crc kubenswrapper[4998]: I1203 16:19:04.209591 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 03 16:19:04 crc kubenswrapper[4998]: I1203 16:19:04.213107 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 03 16:19:04 crc kubenswrapper[4998]: I1203 16:19:04.246252 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zdxk7"] Dec 03 16:19:04 crc kubenswrapper[4998]: I1203 16:19:04.287171 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csm57\" (UniqueName: \"kubernetes.io/projected/8dc87551-eb02-4f3e-b609-1eb49f88a5c6-kube-api-access-csm57\") pod \"openstack-operator-index-zdxk7\" (UID: \"8dc87551-eb02-4f3e-b609-1eb49f88a5c6\") " pod="openstack-operators/openstack-operator-index-zdxk7" Dec 03 16:19:04 crc kubenswrapper[4998]: I1203 16:19:04.388682 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csm57\" (UniqueName: \"kubernetes.io/projected/8dc87551-eb02-4f3e-b609-1eb49f88a5c6-kube-api-access-csm57\") pod \"openstack-operator-index-zdxk7\" (UID: \"8dc87551-eb02-4f3e-b609-1eb49f88a5c6\") " pod="openstack-operators/openstack-operator-index-zdxk7" Dec 03 16:19:04 crc kubenswrapper[4998]: I1203 16:19:04.409564 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csm57\" (UniqueName: \"kubernetes.io/projected/8dc87551-eb02-4f3e-b609-1eb49f88a5c6-kube-api-access-csm57\") pod \"openstack-operator-index-zdxk7\" (UID: \"8dc87551-eb02-4f3e-b609-1eb49f88a5c6\") " pod="openstack-operators/openstack-operator-index-zdxk7" Dec 03 16:19:04 crc kubenswrapper[4998]: I1203 16:19:04.527709 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zdxk7" Dec 03 16:19:04 crc kubenswrapper[4998]: I1203 16:19:04.730020 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zdxk7"] Dec 03 16:19:05 crc kubenswrapper[4998]: I1203 16:19:05.251577 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zdxk7" event={"ID":"8dc87551-eb02-4f3e-b609-1eb49f88a5c6","Type":"ContainerStarted","Data":"e98cf86569f53d002ec82b2b4ca2808d550e58b31d5a1d6cb5a7a7053b276331"} Dec 03 16:19:07 crc kubenswrapper[4998]: I1203 16:19:07.567006 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-zdxk7"] Dec 03 16:19:08 crc kubenswrapper[4998]: I1203 16:19:08.190931 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-8rmlq"] Dec 03 16:19:08 crc kubenswrapper[4998]: I1203 16:19:08.192233 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-8rmlq" Dec 03 16:19:08 crc kubenswrapper[4998]: I1203 16:19:08.204751 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-8rmlq"] Dec 03 16:19:08 crc kubenswrapper[4998]: I1203 16:19:08.243099 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6qnj\" (UniqueName: \"kubernetes.io/projected/3e42935d-e3d2-48a6-a787-caa1ed1f7ef1-kube-api-access-p6qnj\") pod \"openstack-operator-index-8rmlq\" (UID: \"3e42935d-e3d2-48a6-a787-caa1ed1f7ef1\") " pod="openstack-operators/openstack-operator-index-8rmlq" Dec 03 16:19:08 crc kubenswrapper[4998]: I1203 16:19:08.344675 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6qnj\" (UniqueName: \"kubernetes.io/projected/3e42935d-e3d2-48a6-a787-caa1ed1f7ef1-kube-api-access-p6qnj\") pod \"openstack-operator-index-8rmlq\" (UID: \"3e42935d-e3d2-48a6-a787-caa1ed1f7ef1\") " pod="openstack-operators/openstack-operator-index-8rmlq" Dec 03 16:19:08 crc kubenswrapper[4998]: I1203 16:19:08.369485 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6qnj\" (UniqueName: \"kubernetes.io/projected/3e42935d-e3d2-48a6-a787-caa1ed1f7ef1-kube-api-access-p6qnj\") pod \"openstack-operator-index-8rmlq\" (UID: \"3e42935d-e3d2-48a6-a787-caa1ed1f7ef1\") " pod="openstack-operators/openstack-operator-index-8rmlq" Dec 03 16:19:08 crc kubenswrapper[4998]: I1203 16:19:08.529380 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-8rmlq" Dec 03 16:19:10 crc kubenswrapper[4998]: I1203 16:19:10.204719 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-8rmlq"] Dec 03 16:19:10 crc kubenswrapper[4998]: I1203 16:19:10.292962 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-8rmlq" event={"ID":"3e42935d-e3d2-48a6-a787-caa1ed1f7ef1","Type":"ContainerStarted","Data":"cec07f71eba7383dab227ef3fea5392214c795319bce309e9c2cbfd81a597294"} Dec 03 16:19:10 crc kubenswrapper[4998]: I1203 16:19:10.294080 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zdxk7" event={"ID":"8dc87551-eb02-4f3e-b609-1eb49f88a5c6","Type":"ContainerStarted","Data":"8e5f8fa8083e21f63ac5bd08bd79f8f89baf76ce429954a56edbe0819e1b08d7"} Dec 03 16:19:10 crc kubenswrapper[4998]: I1203 16:19:10.294243 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-zdxk7" podUID="8dc87551-eb02-4f3e-b609-1eb49f88a5c6" containerName="registry-server" containerID="cri-o://8e5f8fa8083e21f63ac5bd08bd79f8f89baf76ce429954a56edbe0819e1b08d7" gracePeriod=2 Dec 03 16:19:10 crc kubenswrapper[4998]: I1203 16:19:10.312780 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-zdxk7" podStartSLOduration=1.065624215 podStartE2EDuration="6.312730567s" podCreationTimestamp="2025-12-03 16:19:04 +0000 UTC" firstStartedPulling="2025-12-03 16:19:04.743662771 +0000 UTC m=+923.355362994" lastFinishedPulling="2025-12-03 16:19:09.990769123 +0000 UTC m=+928.602469346" observedRunningTime="2025-12-03 16:19:10.312368579 +0000 UTC m=+928.924068832" watchObservedRunningTime="2025-12-03 16:19:10.312730567 +0000 UTC m=+928.924430820" Dec 03 16:19:10 crc kubenswrapper[4998]: I1203 16:19:10.791081 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zdxk7" Dec 03 16:19:10 crc kubenswrapper[4998]: I1203 16:19:10.880498 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csm57\" (UniqueName: \"kubernetes.io/projected/8dc87551-eb02-4f3e-b609-1eb49f88a5c6-kube-api-access-csm57\") pod \"8dc87551-eb02-4f3e-b609-1eb49f88a5c6\" (UID: \"8dc87551-eb02-4f3e-b609-1eb49f88a5c6\") " Dec 03 16:19:10 crc kubenswrapper[4998]: I1203 16:19:10.889323 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dc87551-eb02-4f3e-b609-1eb49f88a5c6-kube-api-access-csm57" (OuterVolumeSpecName: "kube-api-access-csm57") pod "8dc87551-eb02-4f3e-b609-1eb49f88a5c6" (UID: "8dc87551-eb02-4f3e-b609-1eb49f88a5c6"). InnerVolumeSpecName "kube-api-access-csm57". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:19:10 crc kubenswrapper[4998]: I1203 16:19:10.982518 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csm57\" (UniqueName: \"kubernetes.io/projected/8dc87551-eb02-4f3e-b609-1eb49f88a5c6-kube-api-access-csm57\") on node \"crc\" DevicePath \"\"" Dec 03 16:19:11 crc kubenswrapper[4998]: I1203 16:19:11.300179 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-8rmlq" event={"ID":"3e42935d-e3d2-48a6-a787-caa1ed1f7ef1","Type":"ContainerStarted","Data":"51619fa1f6b07b760dfd00cd15ba6a45656368282341dd053de23ca18dc7b372"} Dec 03 16:19:11 crc kubenswrapper[4998]: I1203 16:19:11.305207 4998 generic.go:334] "Generic (PLEG): container finished" podID="8dc87551-eb02-4f3e-b609-1eb49f88a5c6" containerID="8e5f8fa8083e21f63ac5bd08bd79f8f89baf76ce429954a56edbe0819e1b08d7" exitCode=0 Dec 03 16:19:11 crc kubenswrapper[4998]: I1203 16:19:11.305245 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zdxk7" event={"ID":"8dc87551-eb02-4f3e-b609-1eb49f88a5c6","Type":"ContainerDied","Data":"8e5f8fa8083e21f63ac5bd08bd79f8f89baf76ce429954a56edbe0819e1b08d7"} Dec 03 16:19:11 crc kubenswrapper[4998]: I1203 16:19:11.305270 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zdxk7" event={"ID":"8dc87551-eb02-4f3e-b609-1eb49f88a5c6","Type":"ContainerDied","Data":"e98cf86569f53d002ec82b2b4ca2808d550e58b31d5a1d6cb5a7a7053b276331"} Dec 03 16:19:11 crc kubenswrapper[4998]: I1203 16:19:11.305280 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zdxk7" Dec 03 16:19:11 crc kubenswrapper[4998]: I1203 16:19:11.305293 4998 scope.go:117] "RemoveContainer" containerID="8e5f8fa8083e21f63ac5bd08bd79f8f89baf76ce429954a56edbe0819e1b08d7" Dec 03 16:19:11 crc kubenswrapper[4998]: I1203 16:19:11.320893 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-8rmlq" podStartSLOduration=3.275074008 podStartE2EDuration="3.320867272s" podCreationTimestamp="2025-12-03 16:19:08 +0000 UTC" firstStartedPulling="2025-12-03 16:19:10.221727433 +0000 UTC m=+928.833427656" lastFinishedPulling="2025-12-03 16:19:10.267520657 +0000 UTC m=+928.879220920" observedRunningTime="2025-12-03 16:19:11.314634834 +0000 UTC m=+929.926335067" watchObservedRunningTime="2025-12-03 16:19:11.320867272 +0000 UTC m=+929.932567505" Dec 03 16:19:11 crc kubenswrapper[4998]: I1203 16:19:11.332870 4998 scope.go:117] "RemoveContainer" containerID="8e5f8fa8083e21f63ac5bd08bd79f8f89baf76ce429954a56edbe0819e1b08d7" Dec 03 16:19:11 crc kubenswrapper[4998]: E1203 16:19:11.333240 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e5f8fa8083e21f63ac5bd08bd79f8f89baf76ce429954a56edbe0819e1b08d7\": container with ID starting with 8e5f8fa8083e21f63ac5bd08bd79f8f89baf76ce429954a56edbe0819e1b08d7 not found: ID does not exist" containerID="8e5f8fa8083e21f63ac5bd08bd79f8f89baf76ce429954a56edbe0819e1b08d7" Dec 03 16:19:11 crc kubenswrapper[4998]: I1203 16:19:11.333281 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e5f8fa8083e21f63ac5bd08bd79f8f89baf76ce429954a56edbe0819e1b08d7"} err="failed to get container status \"8e5f8fa8083e21f63ac5bd08bd79f8f89baf76ce429954a56edbe0819e1b08d7\": rpc error: code = NotFound desc = could not find container \"8e5f8fa8083e21f63ac5bd08bd79f8f89baf76ce429954a56edbe0819e1b08d7\": container with ID starting with 8e5f8fa8083e21f63ac5bd08bd79f8f89baf76ce429954a56edbe0819e1b08d7 not found: ID does not exist" Dec 03 16:19:11 crc kubenswrapper[4998]: I1203 16:19:11.343740 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-zdxk7"] Dec 03 16:19:11 crc kubenswrapper[4998]: I1203 16:19:11.347422 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-zdxk7"] Dec 03 16:19:11 crc kubenswrapper[4998]: I1203 16:19:11.695115 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dc87551-eb02-4f3e-b609-1eb49f88a5c6" path="/var/lib/kubelet/pods/8dc87551-eb02-4f3e-b609-1eb49f88a5c6/volumes" Dec 03 16:19:18 crc kubenswrapper[4998]: I1203 16:19:18.530324 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-8rmlq" Dec 03 16:19:18 crc kubenswrapper[4998]: I1203 16:19:18.530839 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-8rmlq" Dec 03 16:19:18 crc kubenswrapper[4998]: I1203 16:19:18.561972 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-8rmlq" Dec 03 16:19:19 crc kubenswrapper[4998]: I1203 16:19:19.412884 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-8rmlq" Dec 03 16:19:20 crc kubenswrapper[4998]: I1203 16:19:20.653884 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl"] Dec 03 16:19:20 crc kubenswrapper[4998]: E1203 16:19:20.655001 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dc87551-eb02-4f3e-b609-1eb49f88a5c6" containerName="registry-server" Dec 03 16:19:20 crc kubenswrapper[4998]: I1203 16:19:20.655113 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dc87551-eb02-4f3e-b609-1eb49f88a5c6" containerName="registry-server" Dec 03 16:19:20 crc kubenswrapper[4998]: I1203 16:19:20.655294 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dc87551-eb02-4f3e-b609-1eb49f88a5c6" containerName="registry-server" Dec 03 16:19:20 crc kubenswrapper[4998]: I1203 16:19:20.656478 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl" Dec 03 16:19:20 crc kubenswrapper[4998]: I1203 16:19:20.662408 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-jsqp2" Dec 03 16:19:20 crc kubenswrapper[4998]: I1203 16:19:20.673775 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl"] Dec 03 16:19:20 crc kubenswrapper[4998]: I1203 16:19:20.748970 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/42863f48-3fe1-4510-a729-cd3c0afa39ec-util\") pod \"6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl\" (UID: \"42863f48-3fe1-4510-a729-cd3c0afa39ec\") " pod="openstack-operators/6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl" Dec 03 16:19:20 crc kubenswrapper[4998]: I1203 16:19:20.749171 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/42863f48-3fe1-4510-a729-cd3c0afa39ec-bundle\") pod \"6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl\" (UID: \"42863f48-3fe1-4510-a729-cd3c0afa39ec\") " pod="openstack-operators/6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl" Dec 03 16:19:20 crc kubenswrapper[4998]: I1203 16:19:20.749292 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n2n7\" (UniqueName: \"kubernetes.io/projected/42863f48-3fe1-4510-a729-cd3c0afa39ec-kube-api-access-9n2n7\") pod \"6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl\" (UID: \"42863f48-3fe1-4510-a729-cd3c0afa39ec\") " pod="openstack-operators/6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl" Dec 03 16:19:20 crc kubenswrapper[4998]: I1203 16:19:20.850549 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/42863f48-3fe1-4510-a729-cd3c0afa39ec-util\") pod \"6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl\" (UID: \"42863f48-3fe1-4510-a729-cd3c0afa39ec\") " pod="openstack-operators/6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl" Dec 03 16:19:20 crc kubenswrapper[4998]: I1203 16:19:20.850632 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/42863f48-3fe1-4510-a729-cd3c0afa39ec-bundle\") pod \"6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl\" (UID: \"42863f48-3fe1-4510-a729-cd3c0afa39ec\") " pod="openstack-operators/6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl" Dec 03 16:19:20 crc kubenswrapper[4998]: I1203 16:19:20.850675 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n2n7\" (UniqueName: \"kubernetes.io/projected/42863f48-3fe1-4510-a729-cd3c0afa39ec-kube-api-access-9n2n7\") pod \"6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl\" (UID: \"42863f48-3fe1-4510-a729-cd3c0afa39ec\") " pod="openstack-operators/6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl" Dec 03 16:19:20 crc kubenswrapper[4998]: I1203 16:19:20.851851 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/42863f48-3fe1-4510-a729-cd3c0afa39ec-bundle\") pod \"6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl\" (UID: \"42863f48-3fe1-4510-a729-cd3c0afa39ec\") " pod="openstack-operators/6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl" Dec 03 16:19:20 crc kubenswrapper[4998]: I1203 16:19:20.852073 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/42863f48-3fe1-4510-a729-cd3c0afa39ec-util\") pod \"6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl\" (UID: \"42863f48-3fe1-4510-a729-cd3c0afa39ec\") " pod="openstack-operators/6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl" Dec 03 16:19:20 crc kubenswrapper[4998]: I1203 16:19:20.880324 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n2n7\" (UniqueName: \"kubernetes.io/projected/42863f48-3fe1-4510-a729-cd3c0afa39ec-kube-api-access-9n2n7\") pod \"6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl\" (UID: \"42863f48-3fe1-4510-a729-cd3c0afa39ec\") " pod="openstack-operators/6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl" Dec 03 16:19:20 crc kubenswrapper[4998]: I1203 16:19:20.983432 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl" Dec 03 16:19:21 crc kubenswrapper[4998]: I1203 16:19:21.270262 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl"] Dec 03 16:19:21 crc kubenswrapper[4998]: I1203 16:19:21.383630 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl" event={"ID":"42863f48-3fe1-4510-a729-cd3c0afa39ec","Type":"ContainerStarted","Data":"d1dd83f302e575b8526a18cea1a291330c0bb567aeb5a451dfed014859ca91ee"} Dec 03 16:19:23 crc kubenswrapper[4998]: I1203 16:19:23.401580 4998 generic.go:334] "Generic (PLEG): container finished" podID="42863f48-3fe1-4510-a729-cd3c0afa39ec" containerID="d27ac4aeacaa2e9532b8524a01d654d53d839b17f12249f15e64f86249a99858" exitCode=0 Dec 03 16:19:23 crc kubenswrapper[4998]: I1203 16:19:23.401654 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl" event={"ID":"42863f48-3fe1-4510-a729-cd3c0afa39ec","Type":"ContainerDied","Data":"d27ac4aeacaa2e9532b8524a01d654d53d839b17f12249f15e64f86249a99858"} Dec 03 16:19:24 crc kubenswrapper[4998]: I1203 16:19:24.415979 4998 generic.go:334] "Generic (PLEG): container finished" podID="42863f48-3fe1-4510-a729-cd3c0afa39ec" containerID="841c3561e2abcec4c6adcfcefcdbed9444158944ec48f45dbdc639b368e2b0eb" exitCode=0 Dec 03 16:19:24 crc kubenswrapper[4998]: I1203 16:19:24.416026 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl" event={"ID":"42863f48-3fe1-4510-a729-cd3c0afa39ec","Type":"ContainerDied","Data":"841c3561e2abcec4c6adcfcefcdbed9444158944ec48f45dbdc639b368e2b0eb"} Dec 03 16:19:25 crc kubenswrapper[4998]: I1203 16:19:25.429581 4998 generic.go:334] "Generic (PLEG): container finished" podID="42863f48-3fe1-4510-a729-cd3c0afa39ec" containerID="99f8e7270f68428e659ae6ddb543e23318b826e39cc471947a70e94f896f1753" exitCode=0 Dec 03 16:19:25 crc kubenswrapper[4998]: I1203 16:19:25.429654 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl" event={"ID":"42863f48-3fe1-4510-a729-cd3c0afa39ec","Type":"ContainerDied","Data":"99f8e7270f68428e659ae6ddb543e23318b826e39cc471947a70e94f896f1753"} Dec 03 16:19:26 crc kubenswrapper[4998]: I1203 16:19:26.716473 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl" Dec 03 16:19:26 crc kubenswrapper[4998]: I1203 16:19:26.841250 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/42863f48-3fe1-4510-a729-cd3c0afa39ec-bundle\") pod \"42863f48-3fe1-4510-a729-cd3c0afa39ec\" (UID: \"42863f48-3fe1-4510-a729-cd3c0afa39ec\") " Dec 03 16:19:26 crc kubenswrapper[4998]: I1203 16:19:26.841337 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9n2n7\" (UniqueName: \"kubernetes.io/projected/42863f48-3fe1-4510-a729-cd3c0afa39ec-kube-api-access-9n2n7\") pod \"42863f48-3fe1-4510-a729-cd3c0afa39ec\" (UID: \"42863f48-3fe1-4510-a729-cd3c0afa39ec\") " Dec 03 16:19:26 crc kubenswrapper[4998]: I1203 16:19:26.841413 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/42863f48-3fe1-4510-a729-cd3c0afa39ec-util\") pod \"42863f48-3fe1-4510-a729-cd3c0afa39ec\" (UID: \"42863f48-3fe1-4510-a729-cd3c0afa39ec\") " Dec 03 16:19:26 crc kubenswrapper[4998]: I1203 16:19:26.842073 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42863f48-3fe1-4510-a729-cd3c0afa39ec-bundle" (OuterVolumeSpecName: "bundle") pod "42863f48-3fe1-4510-a729-cd3c0afa39ec" (UID: "42863f48-3fe1-4510-a729-cd3c0afa39ec"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:19:26 crc kubenswrapper[4998]: I1203 16:19:26.846334 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42863f48-3fe1-4510-a729-cd3c0afa39ec-kube-api-access-9n2n7" (OuterVolumeSpecName: "kube-api-access-9n2n7") pod "42863f48-3fe1-4510-a729-cd3c0afa39ec" (UID: "42863f48-3fe1-4510-a729-cd3c0afa39ec"). InnerVolumeSpecName "kube-api-access-9n2n7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:19:26 crc kubenswrapper[4998]: I1203 16:19:26.854933 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42863f48-3fe1-4510-a729-cd3c0afa39ec-util" (OuterVolumeSpecName: "util") pod "42863f48-3fe1-4510-a729-cd3c0afa39ec" (UID: "42863f48-3fe1-4510-a729-cd3c0afa39ec"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:19:26 crc kubenswrapper[4998]: I1203 16:19:26.942615 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9n2n7\" (UniqueName: \"kubernetes.io/projected/42863f48-3fe1-4510-a729-cd3c0afa39ec-kube-api-access-9n2n7\") on node \"crc\" DevicePath \"\"" Dec 03 16:19:26 crc kubenswrapper[4998]: I1203 16:19:26.942922 4998 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/42863f48-3fe1-4510-a729-cd3c0afa39ec-util\") on node \"crc\" DevicePath \"\"" Dec 03 16:19:26 crc kubenswrapper[4998]: I1203 16:19:26.942935 4998 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/42863f48-3fe1-4510-a729-cd3c0afa39ec-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:19:27 crc kubenswrapper[4998]: I1203 16:19:27.111522 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:19:27 crc kubenswrapper[4998]: I1203 16:19:27.111933 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:19:27 crc kubenswrapper[4998]: I1203 16:19:27.112171 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:19:27 crc kubenswrapper[4998]: I1203 16:19:27.113483 4998 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5b97e0b62564fae4effbf36e0feb1d3ce5f2c4a4d5ac7c086146fd76fc8ee224"} pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 16:19:27 crc kubenswrapper[4998]: I1203 16:19:27.114069 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" containerID="cri-o://5b97e0b62564fae4effbf36e0feb1d3ce5f2c4a4d5ac7c086146fd76fc8ee224" gracePeriod=600 Dec 03 16:19:27 crc kubenswrapper[4998]: I1203 16:19:27.445119 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl" Dec 03 16:19:27 crc kubenswrapper[4998]: I1203 16:19:27.445241 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl" event={"ID":"42863f48-3fe1-4510-a729-cd3c0afa39ec","Type":"ContainerDied","Data":"d1dd83f302e575b8526a18cea1a291330c0bb567aeb5a451dfed014859ca91ee"} Dec 03 16:19:27 crc kubenswrapper[4998]: I1203 16:19:27.445311 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1dd83f302e575b8526a18cea1a291330c0bb567aeb5a451dfed014859ca91ee" Dec 03 16:19:27 crc kubenswrapper[4998]: I1203 16:19:27.448872 4998 generic.go:334] "Generic (PLEG): container finished" podID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerID="5b97e0b62564fae4effbf36e0feb1d3ce5f2c4a4d5ac7c086146fd76fc8ee224" exitCode=0 Dec 03 16:19:27 crc kubenswrapper[4998]: I1203 16:19:27.448938 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerDied","Data":"5b97e0b62564fae4effbf36e0feb1d3ce5f2c4a4d5ac7c086146fd76fc8ee224"} Dec 03 16:19:27 crc kubenswrapper[4998]: I1203 16:19:27.448998 4998 scope.go:117] "RemoveContainer" containerID="bc9f09c91a152bc17df6a34b6b589619602b656e78ba0e9011a364c51abd030e" Dec 03 16:19:28 crc kubenswrapper[4998]: I1203 16:19:28.462863 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"4e819484c40f765e06b9e1c5a211c516f5099460ab695ea890ca1412f47f0680"} Dec 03 16:19:34 crc kubenswrapper[4998]: I1203 16:19:34.639318 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-59449b5fc9-lss2t"] Dec 03 16:19:34 crc kubenswrapper[4998]: E1203 16:19:34.640809 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42863f48-3fe1-4510-a729-cd3c0afa39ec" containerName="pull" Dec 03 16:19:34 crc kubenswrapper[4998]: I1203 16:19:34.640899 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="42863f48-3fe1-4510-a729-cd3c0afa39ec" containerName="pull" Dec 03 16:19:34 crc kubenswrapper[4998]: E1203 16:19:34.641003 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42863f48-3fe1-4510-a729-cd3c0afa39ec" containerName="extract" Dec 03 16:19:34 crc kubenswrapper[4998]: I1203 16:19:34.641087 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="42863f48-3fe1-4510-a729-cd3c0afa39ec" containerName="extract" Dec 03 16:19:34 crc kubenswrapper[4998]: E1203 16:19:34.641163 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42863f48-3fe1-4510-a729-cd3c0afa39ec" containerName="util" Dec 03 16:19:34 crc kubenswrapper[4998]: I1203 16:19:34.641232 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="42863f48-3fe1-4510-a729-cd3c0afa39ec" containerName="util" Dec 03 16:19:34 crc kubenswrapper[4998]: I1203 16:19:34.641422 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="42863f48-3fe1-4510-a729-cd3c0afa39ec" containerName="extract" Dec 03 16:19:34 crc kubenswrapper[4998]: I1203 16:19:34.642004 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-59449b5fc9-lss2t" Dec 03 16:19:34 crc kubenswrapper[4998]: I1203 16:19:34.644455 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-xnxcd" Dec 03 16:19:34 crc kubenswrapper[4998]: I1203 16:19:34.667628 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-59449b5fc9-lss2t"] Dec 03 16:19:34 crc kubenswrapper[4998]: I1203 16:19:34.750950 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqww2\" (UniqueName: \"kubernetes.io/projected/f730843a-f6a7-4e56-bd88-333d2531d2f0-kube-api-access-jqww2\") pod \"openstack-operator-controller-operator-59449b5fc9-lss2t\" (UID: \"f730843a-f6a7-4e56-bd88-333d2531d2f0\") " pod="openstack-operators/openstack-operator-controller-operator-59449b5fc9-lss2t" Dec 03 16:19:34 crc kubenswrapper[4998]: I1203 16:19:34.852145 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqww2\" (UniqueName: \"kubernetes.io/projected/f730843a-f6a7-4e56-bd88-333d2531d2f0-kube-api-access-jqww2\") pod \"openstack-operator-controller-operator-59449b5fc9-lss2t\" (UID: \"f730843a-f6a7-4e56-bd88-333d2531d2f0\") " pod="openstack-operators/openstack-operator-controller-operator-59449b5fc9-lss2t" Dec 03 16:19:34 crc kubenswrapper[4998]: I1203 16:19:34.869419 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqww2\" (UniqueName: \"kubernetes.io/projected/f730843a-f6a7-4e56-bd88-333d2531d2f0-kube-api-access-jqww2\") pod \"openstack-operator-controller-operator-59449b5fc9-lss2t\" (UID: \"f730843a-f6a7-4e56-bd88-333d2531d2f0\") " pod="openstack-operators/openstack-operator-controller-operator-59449b5fc9-lss2t" Dec 03 16:19:34 crc kubenswrapper[4998]: I1203 16:19:34.967113 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-59449b5fc9-lss2t" Dec 03 16:19:35 crc kubenswrapper[4998]: I1203 16:19:35.443677 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-59449b5fc9-lss2t"] Dec 03 16:19:35 crc kubenswrapper[4998]: I1203 16:19:35.511981 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-59449b5fc9-lss2t" event={"ID":"f730843a-f6a7-4e56-bd88-333d2531d2f0","Type":"ContainerStarted","Data":"2c67d246ea09c364b1128b13322b843e4be03786183775dc8567cfcb6f990e1b"} Dec 03 16:19:39 crc kubenswrapper[4998]: I1203 16:19:39.540563 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-59449b5fc9-lss2t" event={"ID":"f730843a-f6a7-4e56-bd88-333d2531d2f0","Type":"ContainerStarted","Data":"ea3bab2f9948916e35a49da938b52be5551150607a7c297d1bf3da3847ab2299"} Dec 03 16:19:39 crc kubenswrapper[4998]: I1203 16:19:39.541225 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-59449b5fc9-lss2t" Dec 03 16:19:39 crc kubenswrapper[4998]: I1203 16:19:39.579933 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-59449b5fc9-lss2t" podStartSLOduration=1.823911672 podStartE2EDuration="5.579913355s" podCreationTimestamp="2025-12-03 16:19:34 +0000 UTC" firstStartedPulling="2025-12-03 16:19:35.453820568 +0000 UTC m=+954.065520791" lastFinishedPulling="2025-12-03 16:19:39.209822251 +0000 UTC m=+957.821522474" observedRunningTime="2025-12-03 16:19:39.57537023 +0000 UTC m=+958.187070453" watchObservedRunningTime="2025-12-03 16:19:39.579913355 +0000 UTC m=+958.191613578" Dec 03 16:19:44 crc kubenswrapper[4998]: I1203 16:19:44.969979 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-59449b5fc9-lss2t" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.108746 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-5h65w"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.111523 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5h65w" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.114232 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-bf56m" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.119732 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-5h65w"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.141918 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-tv6bd"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.143220 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-tv6bd" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.146272 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-xr69z" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.151292 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-tv6bd"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.175729 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-m4gc7"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.176895 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-m4gc7" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.179157 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-6nlt7" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.184532 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-tkz2j"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.185568 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-tkz2j" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.189797 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-j7lnw" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.194699 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-m4gc7"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.216661 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-tkz2j"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.234008 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d2gcr"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.235036 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d2gcr" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.236966 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-m7trz" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.240091 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gzg25"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.241042 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gzg25" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.242721 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-gjk68" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.274859 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d2gcr"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.280441 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-hr74g"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.281360 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-hr74g" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.283728 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.290221 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-8h8gq" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.307447 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87l7h\" (UniqueName: \"kubernetes.io/projected/a67cb778-507e-4dba-9ae4-60ae60315a0a-kube-api-access-87l7h\") pod \"cinder-operator-controller-manager-859b6ccc6-tv6bd\" (UID: \"a67cb778-507e-4dba-9ae4-60ae60315a0a\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-tv6bd" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.307488 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlllz\" (UniqueName: \"kubernetes.io/projected/624bac01-c09f-424f-9000-feb138765cc4-kube-api-access-zlllz\") pod \"glance-operator-controller-manager-77987cd8cd-tkz2j\" (UID: \"624bac01-c09f-424f-9000-feb138765cc4\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-tkz2j" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.307527 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksr2r\" (UniqueName: \"kubernetes.io/projected/67e57367-12f9-487d-b691-ca243a28b056-kube-api-access-ksr2r\") pod \"designate-operator-controller-manager-78b4bc895b-m4gc7\" (UID: \"67e57367-12f9-487d-b691-ca243a28b056\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-m4gc7" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.307684 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjrsz\" (UniqueName: \"kubernetes.io/projected/da7a11af-1959-4c47-96d3-4c5f4c469493-kube-api-access-fjrsz\") pod \"barbican-operator-controller-manager-7d9dfd778-5h65w\" (UID: \"da7a11af-1959-4c47-96d3-4c5f4c469493\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5h65w" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.312343 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gzg25"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.330678 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-tcltm"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.332014 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-tcltm" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.335071 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-n5g9d" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.341457 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-hr74g"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.350526 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-tcltm"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.369109 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-mn69l"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.370369 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-mn69l" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.377669 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-99mtk" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.386594 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-nfm6p"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.387631 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nfm6p" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.393204 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-tz4ct" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.406548 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-nfm6p"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.409989 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhjfb\" (UniqueName: \"kubernetes.io/projected/dcb7317b-497a-4135-be25-3462a91406e9-kube-api-access-qhjfb\") pod \"keystone-operator-controller-manager-7765d96ddf-nfm6p\" (UID: \"dcb7317b-497a-4135-be25-3462a91406e9\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nfm6p" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.410027 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgvv9\" (UniqueName: \"kubernetes.io/projected/ade59116-1aaf-44e0-b47b-15bdc7b402f0-kube-api-access-cgvv9\") pod \"infra-operator-controller-manager-57548d458d-hr74g\" (UID: \"ade59116-1aaf-44e0-b47b-15bdc7b402f0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-hr74g" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.410051 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87l7h\" (UniqueName: \"kubernetes.io/projected/a67cb778-507e-4dba-9ae4-60ae60315a0a-kube-api-access-87l7h\") pod \"cinder-operator-controller-manager-859b6ccc6-tv6bd\" (UID: \"a67cb778-507e-4dba-9ae4-60ae60315a0a\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-tv6bd" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.410071 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlllz\" (UniqueName: \"kubernetes.io/projected/624bac01-c09f-424f-9000-feb138765cc4-kube-api-access-zlllz\") pod \"glance-operator-controller-manager-77987cd8cd-tkz2j\" (UID: \"624bac01-c09f-424f-9000-feb138765cc4\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-tkz2j" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.410098 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plmts\" (UniqueName: \"kubernetes.io/projected/7c372d0c-4cac-4e4c-bc8f-409b5c151d53-kube-api-access-plmts\") pod \"horizon-operator-controller-manager-68c6d99b8f-gzg25\" (UID: \"7c372d0c-4cac-4e4c-bc8f-409b5c151d53\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gzg25" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.410118 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nkng\" (UniqueName: \"kubernetes.io/projected/e52d7801-a5c9-47a3-bedc-27174cfbd993-kube-api-access-8nkng\") pod \"heat-operator-controller-manager-5f64f6f8bb-d2gcr\" (UID: \"e52d7801-a5c9-47a3-bedc-27174cfbd993\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d2gcr" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.410134 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksr2r\" (UniqueName: \"kubernetes.io/projected/67e57367-12f9-487d-b691-ca243a28b056-kube-api-access-ksr2r\") pod \"designate-operator-controller-manager-78b4bc895b-m4gc7\" (UID: \"67e57367-12f9-487d-b691-ca243a28b056\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-m4gc7" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.410177 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k8kc\" (UniqueName: \"kubernetes.io/projected/c92de5b9-8ef5-4f4c-b881-82c5f05cd853-kube-api-access-7k8kc\") pod \"manila-operator-controller-manager-7c79b5df47-mn69l\" (UID: \"c92de5b9-8ef5-4f4c-b881-82c5f05cd853\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-mn69l" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.410197 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjrsz\" (UniqueName: \"kubernetes.io/projected/da7a11af-1959-4c47-96d3-4c5f4c469493-kube-api-access-fjrsz\") pod \"barbican-operator-controller-manager-7d9dfd778-5h65w\" (UID: \"da7a11af-1959-4c47-96d3-4c5f4c469493\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5h65w" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.410214 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrmlx\" (UniqueName: \"kubernetes.io/projected/17cdd58c-7a0c-42e0-bcbe-4545050637e3-kube-api-access-rrmlx\") pod \"ironic-operator-controller-manager-6c548fd776-tcltm\" (UID: \"17cdd58c-7a0c-42e0-bcbe-4545050637e3\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-tcltm" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.410241 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ade59116-1aaf-44e0-b47b-15bdc7b402f0-cert\") pod \"infra-operator-controller-manager-57548d458d-hr74g\" (UID: \"ade59116-1aaf-44e0-b47b-15bdc7b402f0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-hr74g" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.468338 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlllz\" (UniqueName: \"kubernetes.io/projected/624bac01-c09f-424f-9000-feb138765cc4-kube-api-access-zlllz\") pod \"glance-operator-controller-manager-77987cd8cd-tkz2j\" (UID: \"624bac01-c09f-424f-9000-feb138765cc4\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-tkz2j" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.468688 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-mn69l"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.508666 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87l7h\" (UniqueName: \"kubernetes.io/projected/a67cb778-507e-4dba-9ae4-60ae60315a0a-kube-api-access-87l7h\") pod \"cinder-operator-controller-manager-859b6ccc6-tv6bd\" (UID: \"a67cb778-507e-4dba-9ae4-60ae60315a0a\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-tv6bd" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.514771 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjrsz\" (UniqueName: \"kubernetes.io/projected/da7a11af-1959-4c47-96d3-4c5f4c469493-kube-api-access-fjrsz\") pod \"barbican-operator-controller-manager-7d9dfd778-5h65w\" (UID: \"da7a11af-1959-4c47-96d3-4c5f4c469493\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5h65w" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.523236 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-tkz2j" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.523913 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k8kc\" (UniqueName: \"kubernetes.io/projected/c92de5b9-8ef5-4f4c-b881-82c5f05cd853-kube-api-access-7k8kc\") pod \"manila-operator-controller-manager-7c79b5df47-mn69l\" (UID: \"c92de5b9-8ef5-4f4c-b881-82c5f05cd853\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-mn69l" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.523917 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksr2r\" (UniqueName: \"kubernetes.io/projected/67e57367-12f9-487d-b691-ca243a28b056-kube-api-access-ksr2r\") pod \"designate-operator-controller-manager-78b4bc895b-m4gc7\" (UID: \"67e57367-12f9-487d-b691-ca243a28b056\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-m4gc7" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.524050 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrmlx\" (UniqueName: \"kubernetes.io/projected/17cdd58c-7a0c-42e0-bcbe-4545050637e3-kube-api-access-rrmlx\") pod \"ironic-operator-controller-manager-6c548fd776-tcltm\" (UID: \"17cdd58c-7a0c-42e0-bcbe-4545050637e3\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-tcltm" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.524096 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ade59116-1aaf-44e0-b47b-15bdc7b402f0-cert\") pod \"infra-operator-controller-manager-57548d458d-hr74g\" (UID: \"ade59116-1aaf-44e0-b47b-15bdc7b402f0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-hr74g" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.524211 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhjfb\" (UniqueName: \"kubernetes.io/projected/dcb7317b-497a-4135-be25-3462a91406e9-kube-api-access-qhjfb\") pod \"keystone-operator-controller-manager-7765d96ddf-nfm6p\" (UID: \"dcb7317b-497a-4135-be25-3462a91406e9\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nfm6p" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.524252 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgvv9\" (UniqueName: \"kubernetes.io/projected/ade59116-1aaf-44e0-b47b-15bdc7b402f0-kube-api-access-cgvv9\") pod \"infra-operator-controller-manager-57548d458d-hr74g\" (UID: \"ade59116-1aaf-44e0-b47b-15bdc7b402f0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-hr74g" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.524864 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plmts\" (UniqueName: \"kubernetes.io/projected/7c372d0c-4cac-4e4c-bc8f-409b5c151d53-kube-api-access-plmts\") pod \"horizon-operator-controller-manager-68c6d99b8f-gzg25\" (UID: \"7c372d0c-4cac-4e4c-bc8f-409b5c151d53\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gzg25" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.524917 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nkng\" (UniqueName: \"kubernetes.io/projected/e52d7801-a5c9-47a3-bedc-27174cfbd993-kube-api-access-8nkng\") pod \"heat-operator-controller-manager-5f64f6f8bb-d2gcr\" (UID: \"e52d7801-a5c9-47a3-bedc-27174cfbd993\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d2gcr" Dec 03 16:20:05 crc kubenswrapper[4998]: E1203 16:20:05.524214 4998 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 16:20:05 crc kubenswrapper[4998]: E1203 16:20:05.525173 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ade59116-1aaf-44e0-b47b-15bdc7b402f0-cert podName:ade59116-1aaf-44e0-b47b-15bdc7b402f0 nodeName:}" failed. No retries permitted until 2025-12-03 16:20:06.025144743 +0000 UTC m=+984.636844966 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ade59116-1aaf-44e0-b47b-15bdc7b402f0-cert") pod "infra-operator-controller-manager-57548d458d-hr74g" (UID: "ade59116-1aaf-44e0-b47b-15bdc7b402f0") : secret "infra-operator-webhook-server-cert" not found Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.552071 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f4jqc"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.555734 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f4jqc" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.572566 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-zbfjp" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.585432 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhjfb\" (UniqueName: \"kubernetes.io/projected/dcb7317b-497a-4135-be25-3462a91406e9-kube-api-access-qhjfb\") pod \"keystone-operator-controller-manager-7765d96ddf-nfm6p\" (UID: \"dcb7317b-497a-4135-be25-3462a91406e9\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nfm6p" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.586153 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k8kc\" (UniqueName: \"kubernetes.io/projected/c92de5b9-8ef5-4f4c-b881-82c5f05cd853-kube-api-access-7k8kc\") pod \"manila-operator-controller-manager-7c79b5df47-mn69l\" (UID: \"c92de5b9-8ef5-4f4c-b881-82c5f05cd853\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-mn69l" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.586623 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nkng\" (UniqueName: \"kubernetes.io/projected/e52d7801-a5c9-47a3-bedc-27174cfbd993-kube-api-access-8nkng\") pod \"heat-operator-controller-manager-5f64f6f8bb-d2gcr\" (UID: \"e52d7801-a5c9-47a3-bedc-27174cfbd993\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d2gcr" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.587587 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgvv9\" (UniqueName: \"kubernetes.io/projected/ade59116-1aaf-44e0-b47b-15bdc7b402f0-kube-api-access-cgvv9\") pod \"infra-operator-controller-manager-57548d458d-hr74g\" (UID: \"ade59116-1aaf-44e0-b47b-15bdc7b402f0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-hr74g" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.589129 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrmlx\" (UniqueName: \"kubernetes.io/projected/17cdd58c-7a0c-42e0-bcbe-4545050637e3-kube-api-access-rrmlx\") pod \"ironic-operator-controller-manager-6c548fd776-tcltm\" (UID: \"17cdd58c-7a0c-42e0-bcbe-4545050637e3\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-tcltm" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.590672 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plmts\" (UniqueName: \"kubernetes.io/projected/7c372d0c-4cac-4e4c-bc8f-409b5c151d53-kube-api-access-plmts\") pod \"horizon-operator-controller-manager-68c6d99b8f-gzg25\" (UID: \"7c372d0c-4cac-4e4c-bc8f-409b5c151d53\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gzg25" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.602886 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f4jqc"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.626092 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmdr6\" (UniqueName: \"kubernetes.io/projected/0c9761a8-29fb-4d99-9bc0-914a041b36ac-kube-api-access-vmdr6\") pod \"mariadb-operator-controller-manager-56bbcc9d85-f4jqc\" (UID: \"0c9761a8-29fb-4d99-9bc0-914a041b36ac\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f4jqc" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.629352 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hzl7f"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.630599 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hzl7f" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.633434 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-jglth" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.635981 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-2h2m5"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.637308 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2h2m5" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.639769 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-p7d5l" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.647902 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hzl7f"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.656227 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-ph8r8"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.657321 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-ph8r8" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.658129 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-tcltm" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.658738 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-qngql" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.699377 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-mn69l" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.714061 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-ph8r8"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.714111 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-2h2m5"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.715431 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.717566 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.724980 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-h5mgr" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.725879 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.727016 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttf7z\" (UniqueName: \"kubernetes.io/projected/0e92e441-8150-4817-8039-9cdc8424e562-kube-api-access-ttf7z\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-hzl7f\" (UID: \"0e92e441-8150-4817-8039-9cdc8424e562\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hzl7f" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.727132 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42a2439a-98ee-4172-813c-9ce7109fb1c0-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz\" (UID: \"42a2439a-98ee-4172-813c-9ce7109fb1c0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.727240 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmdr6\" (UniqueName: \"kubernetes.io/projected/0c9761a8-29fb-4d99-9bc0-914a041b36ac-kube-api-access-vmdr6\") pod \"mariadb-operator-controller-manager-56bbcc9d85-f4jqc\" (UID: \"0c9761a8-29fb-4d99-9bc0-914a041b36ac\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f4jqc" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.727276 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qvr2\" (UniqueName: \"kubernetes.io/projected/42a2439a-98ee-4172-813c-9ce7109fb1c0-kube-api-access-7qvr2\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz\" (UID: \"42a2439a-98ee-4172-813c-9ce7109fb1c0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.727306 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjj8q\" (UniqueName: \"kubernetes.io/projected/f23eb2b1-107d-4167-a695-4479df0ae92a-kube-api-access-wjj8q\") pod \"nova-operator-controller-manager-697bc559fc-2h2m5\" (UID: \"f23eb2b1-107d-4167-a695-4479df0ae92a\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2h2m5" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.727379 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwwrh\" (UniqueName: \"kubernetes.io/projected/d926336a-658f-49bd-81b9-8086c8bb97a5-kube-api-access-dwwrh\") pod \"octavia-operator-controller-manager-998648c74-ph8r8\" (UID: \"d926336a-658f-49bd-81b9-8086c8bb97a5\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-ph8r8" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.731149 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nfm6p" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.731250 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5h65w" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.754673 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-28xct"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.760243 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-tv6bd" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.760549 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-28xct" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.764970 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-tsrft" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.768342 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmdr6\" (UniqueName: \"kubernetes.io/projected/0c9761a8-29fb-4d99-9bc0-914a041b36ac-kube-api-access-vmdr6\") pod \"mariadb-operator-controller-manager-56bbcc9d85-f4jqc\" (UID: \"0c9761a8-29fb-4d99-9bc0-914a041b36ac\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f4jqc" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.773464 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.789353 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-h72w5"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.795702 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-h72w5" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.801556 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-hx9tf" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.803101 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-h72w5"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.808474 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-m4gc7" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.829674 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42a2439a-98ee-4172-813c-9ce7109fb1c0-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz\" (UID: \"42a2439a-98ee-4172-813c-9ce7109fb1c0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.829737 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qvr2\" (UniqueName: \"kubernetes.io/projected/42a2439a-98ee-4172-813c-9ce7109fb1c0-kube-api-access-7qvr2\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz\" (UID: \"42a2439a-98ee-4172-813c-9ce7109fb1c0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.829776 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjj8q\" (UniqueName: \"kubernetes.io/projected/f23eb2b1-107d-4167-a695-4479df0ae92a-kube-api-access-wjj8q\") pod \"nova-operator-controller-manager-697bc559fc-2h2m5\" (UID: \"f23eb2b1-107d-4167-a695-4479df0ae92a\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2h2m5" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.829814 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwwrh\" (UniqueName: \"kubernetes.io/projected/d926336a-658f-49bd-81b9-8086c8bb97a5-kube-api-access-dwwrh\") pod \"octavia-operator-controller-manager-998648c74-ph8r8\" (UID: \"d926336a-658f-49bd-81b9-8086c8bb97a5\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-ph8r8" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.829850 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttf7z\" (UniqueName: \"kubernetes.io/projected/0e92e441-8150-4817-8039-9cdc8424e562-kube-api-access-ttf7z\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-hzl7f\" (UID: \"0e92e441-8150-4817-8039-9cdc8424e562\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hzl7f" Dec 03 16:20:05 crc kubenswrapper[4998]: E1203 16:20:05.830458 4998 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 16:20:05 crc kubenswrapper[4998]: E1203 16:20:05.830498 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/42a2439a-98ee-4172-813c-9ce7109fb1c0-cert podName:42a2439a-98ee-4172-813c-9ce7109fb1c0 nodeName:}" failed. No retries permitted until 2025-12-03 16:20:06.330485788 +0000 UTC m=+984.942186011 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/42a2439a-98ee-4172-813c-9ce7109fb1c0-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" (UID: "42a2439a-98ee-4172-813c-9ce7109fb1c0") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.832386 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-28xct"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.841832 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2wrqc"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.843149 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2wrqc" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.846327 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-4jm2j" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.854002 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d2gcr" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.854274 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttf7z\" (UniqueName: \"kubernetes.io/projected/0e92e441-8150-4817-8039-9cdc8424e562-kube-api-access-ttf7z\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-hzl7f\" (UID: \"0e92e441-8150-4817-8039-9cdc8424e562\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hzl7f" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.860995 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjj8q\" (UniqueName: \"kubernetes.io/projected/f23eb2b1-107d-4167-a695-4479df0ae92a-kube-api-access-wjj8q\") pod \"nova-operator-controller-manager-697bc559fc-2h2m5\" (UID: \"f23eb2b1-107d-4167-a695-4479df0ae92a\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2h2m5" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.862820 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-f6vdv"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.864026 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-f6vdv" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.868440 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-8fpmp" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.869136 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwwrh\" (UniqueName: \"kubernetes.io/projected/d926336a-658f-49bd-81b9-8086c8bb97a5-kube-api-access-dwwrh\") pod \"octavia-operator-controller-manager-998648c74-ph8r8\" (UID: \"d926336a-658f-49bd-81b9-8086c8bb97a5\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-ph8r8" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.873283 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qvr2\" (UniqueName: \"kubernetes.io/projected/42a2439a-98ee-4172-813c-9ce7109fb1c0-kube-api-access-7qvr2\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz\" (UID: \"42a2439a-98ee-4172-813c-9ce7109fb1c0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.876190 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gzg25" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.892701 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2wrqc"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.905301 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-f6vdv"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.909456 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f4jqc" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.930945 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcx75\" (UniqueName: \"kubernetes.io/projected/79a1881c-77c7-4c5c-b2fc-b2d7d9c31496-kube-api-access-lcx75\") pod \"ovn-operator-controller-manager-b6456fdb6-28xct\" (UID: \"79a1881c-77c7-4c5c-b2fc-b2d7d9c31496\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-28xct" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.931074 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsv2h\" (UniqueName: \"kubernetes.io/projected/77fb606c-dea3-4e34-9f4c-5bb305e3b7f4-kube-api-access-rsv2h\") pod \"placement-operator-controller-manager-78f8948974-h72w5\" (UID: \"77fb606c-dea3-4e34-9f4c-5bb305e3b7f4\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-h72w5" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.936914 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-z8g4g"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.938219 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z8g4g" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.941910 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-q9sj2" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.948160 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-z8g4g"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.960856 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hzl7f" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.981385 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-f77f967dc-ndj9c"] Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.983567 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2h2m5" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.984941 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-f77f967dc-ndj9c" Dec 03 16:20:05 crc kubenswrapper[4998]: I1203 16:20:05.987239 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-bgv2d" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.000823 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-f77f967dc-ndj9c"] Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.009470 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-ph8r8" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.034977 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjjsx\" (UniqueName: \"kubernetes.io/projected/d7adeced-3036-49a7-8806-fdbefce6406f-kube-api-access-pjjsx\") pod \"swift-operator-controller-manager-5f8c65bbfc-2wrqc\" (UID: \"d7adeced-3036-49a7-8806-fdbefce6406f\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2wrqc" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.035471 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ade59116-1aaf-44e0-b47b-15bdc7b402f0-cert\") pod \"infra-operator-controller-manager-57548d458d-hr74g\" (UID: \"ade59116-1aaf-44e0-b47b-15bdc7b402f0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-hr74g" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.035533 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsv2h\" (UniqueName: \"kubernetes.io/projected/77fb606c-dea3-4e34-9f4c-5bb305e3b7f4-kube-api-access-rsv2h\") pod \"placement-operator-controller-manager-78f8948974-h72w5\" (UID: \"77fb606c-dea3-4e34-9f4c-5bb305e3b7f4\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-h72w5" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.035567 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcx75\" (UniqueName: \"kubernetes.io/projected/79a1881c-77c7-4c5c-b2fc-b2d7d9c31496-kube-api-access-lcx75\") pod \"ovn-operator-controller-manager-b6456fdb6-28xct\" (UID: \"79a1881c-77c7-4c5c-b2fc-b2d7d9c31496\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-28xct" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.035586 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gcmg\" (UniqueName: \"kubernetes.io/projected/aa3ac0c7-ea42-4728-8406-0ce3f9a96c21-kube-api-access-6gcmg\") pod \"telemetry-operator-controller-manager-76cc84c6bb-f6vdv\" (UID: \"aa3ac0c7-ea42-4728-8406-0ce3f9a96c21\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-f6vdv" Dec 03 16:20:06 crc kubenswrapper[4998]: E1203 16:20:06.035743 4998 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 16:20:06 crc kubenswrapper[4998]: E1203 16:20:06.035802 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ade59116-1aaf-44e0-b47b-15bdc7b402f0-cert podName:ade59116-1aaf-44e0-b47b-15bdc7b402f0 nodeName:}" failed. No retries permitted until 2025-12-03 16:20:07.035788537 +0000 UTC m=+985.647488750 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ade59116-1aaf-44e0-b47b-15bdc7b402f0-cert") pod "infra-operator-controller-manager-57548d458d-hr74g" (UID: "ade59116-1aaf-44e0-b47b-15bdc7b402f0") : secret "infra-operator-webhook-server-cert" not found Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.052693 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd"] Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.053600 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.054883 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsv2h\" (UniqueName: \"kubernetes.io/projected/77fb606c-dea3-4e34-9f4c-5bb305e3b7f4-kube-api-access-rsv2h\") pod \"placement-operator-controller-manager-78f8948974-h72w5\" (UID: \"77fb606c-dea3-4e34-9f4c-5bb305e3b7f4\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-h72w5" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.057101 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd"] Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.059573 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.059783 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.060559 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-vdpph" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.076683 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fqdqg"] Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.077251 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcx75\" (UniqueName: \"kubernetes.io/projected/79a1881c-77c7-4c5c-b2fc-b2d7d9c31496-kube-api-access-lcx75\") pod \"ovn-operator-controller-manager-b6456fdb6-28xct\" (UID: \"79a1881c-77c7-4c5c-b2fc-b2d7d9c31496\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-28xct" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.077575 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fqdqg" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.080316 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-bcf5r" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.094124 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-28xct" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.094939 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fqdqg"] Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.119657 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-h72w5" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.124379 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-tkz2j"] Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.137301 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkd5c\" (UniqueName: \"kubernetes.io/projected/6b89567b-4963-4113-9900-a132d873d260-kube-api-access-gkd5c\") pod \"test-operator-controller-manager-5854674fcc-z8g4g\" (UID: \"6b89567b-4963-4113-9900-a132d873d260\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-z8g4g" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.137351 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gcmg\" (UniqueName: \"kubernetes.io/projected/aa3ac0c7-ea42-4728-8406-0ce3f9a96c21-kube-api-access-6gcmg\") pod \"telemetry-operator-controller-manager-76cc84c6bb-f6vdv\" (UID: \"aa3ac0c7-ea42-4728-8406-0ce3f9a96c21\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-f6vdv" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.137388 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjjsx\" (UniqueName: \"kubernetes.io/projected/d7adeced-3036-49a7-8806-fdbefce6406f-kube-api-access-pjjsx\") pod \"swift-operator-controller-manager-5f8c65bbfc-2wrqc\" (UID: \"d7adeced-3036-49a7-8806-fdbefce6406f\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2wrqc" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.137434 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrd78\" (UniqueName: \"kubernetes.io/projected/ac213ee6-009b-4098-9d18-6dfc8cb97a0e-kube-api-access-xrd78\") pod \"watcher-operator-controller-manager-f77f967dc-ndj9c\" (UID: \"ac213ee6-009b-4098-9d18-6dfc8cb97a0e\") " pod="openstack-operators/watcher-operator-controller-manager-f77f967dc-ndj9c" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.154657 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjjsx\" (UniqueName: \"kubernetes.io/projected/d7adeced-3036-49a7-8806-fdbefce6406f-kube-api-access-pjjsx\") pod \"swift-operator-controller-manager-5f8c65bbfc-2wrqc\" (UID: \"d7adeced-3036-49a7-8806-fdbefce6406f\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2wrqc" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.158923 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gcmg\" (UniqueName: \"kubernetes.io/projected/aa3ac0c7-ea42-4728-8406-0ce3f9a96c21-kube-api-access-6gcmg\") pod \"telemetry-operator-controller-manager-76cc84c6bb-f6vdv\" (UID: \"aa3ac0c7-ea42-4728-8406-0ce3f9a96c21\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-f6vdv" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.183834 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2wrqc" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.184649 4998 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.200471 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-f6vdv" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.238518 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrd78\" (UniqueName: \"kubernetes.io/projected/ac213ee6-009b-4098-9d18-6dfc8cb97a0e-kube-api-access-xrd78\") pod \"watcher-operator-controller-manager-f77f967dc-ndj9c\" (UID: \"ac213ee6-009b-4098-9d18-6dfc8cb97a0e\") " pod="openstack-operators/watcher-operator-controller-manager-f77f967dc-ndj9c" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.238554 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjnhc\" (UniqueName: \"kubernetes.io/projected/9f587468-9ae1-43af-8652-f4394865ff37-kube-api-access-vjnhc\") pod \"rabbitmq-cluster-operator-manager-668c99d594-fqdqg\" (UID: \"9f587468-9ae1-43af-8652-f4394865ff37\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fqdqg" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.238610 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-webhook-certs\") pod \"openstack-operator-controller-manager-67847ddbd8-t8wbd\" (UID: \"07d50ab9-95b8-4c58-9192-445ec4e32b59\") " pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.238634 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkd5c\" (UniqueName: \"kubernetes.io/projected/6b89567b-4963-4113-9900-a132d873d260-kube-api-access-gkd5c\") pod \"test-operator-controller-manager-5854674fcc-z8g4g\" (UID: \"6b89567b-4963-4113-9900-a132d873d260\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-z8g4g" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.241512 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-metrics-certs\") pod \"openstack-operator-controller-manager-67847ddbd8-t8wbd\" (UID: \"07d50ab9-95b8-4c58-9192-445ec4e32b59\") " pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.241600 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szw9p\" (UniqueName: \"kubernetes.io/projected/07d50ab9-95b8-4c58-9192-445ec4e32b59-kube-api-access-szw9p\") pod \"openstack-operator-controller-manager-67847ddbd8-t8wbd\" (UID: \"07d50ab9-95b8-4c58-9192-445ec4e32b59\") " pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.257681 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrd78\" (UniqueName: \"kubernetes.io/projected/ac213ee6-009b-4098-9d18-6dfc8cb97a0e-kube-api-access-xrd78\") pod \"watcher-operator-controller-manager-f77f967dc-ndj9c\" (UID: \"ac213ee6-009b-4098-9d18-6dfc8cb97a0e\") " pod="openstack-operators/watcher-operator-controller-manager-f77f967dc-ndj9c" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.262127 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkd5c\" (UniqueName: \"kubernetes.io/projected/6b89567b-4963-4113-9900-a132d873d260-kube-api-access-gkd5c\") pod \"test-operator-controller-manager-5854674fcc-z8g4g\" (UID: \"6b89567b-4963-4113-9900-a132d873d260\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-z8g4g" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.272879 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-tcltm"] Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.319218 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z8g4g" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.339683 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-f77f967dc-ndj9c" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.342623 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42a2439a-98ee-4172-813c-9ce7109fb1c0-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz\" (UID: \"42a2439a-98ee-4172-813c-9ce7109fb1c0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.342660 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-webhook-certs\") pod \"openstack-operator-controller-manager-67847ddbd8-t8wbd\" (UID: \"07d50ab9-95b8-4c58-9192-445ec4e32b59\") " pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.342693 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-metrics-certs\") pod \"openstack-operator-controller-manager-67847ddbd8-t8wbd\" (UID: \"07d50ab9-95b8-4c58-9192-445ec4e32b59\") " pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.342725 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szw9p\" (UniqueName: \"kubernetes.io/projected/07d50ab9-95b8-4c58-9192-445ec4e32b59-kube-api-access-szw9p\") pod \"openstack-operator-controller-manager-67847ddbd8-t8wbd\" (UID: \"07d50ab9-95b8-4c58-9192-445ec4e32b59\") " pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.342793 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjnhc\" (UniqueName: \"kubernetes.io/projected/9f587468-9ae1-43af-8652-f4394865ff37-kube-api-access-vjnhc\") pod \"rabbitmq-cluster-operator-manager-668c99d594-fqdqg\" (UID: \"9f587468-9ae1-43af-8652-f4394865ff37\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fqdqg" Dec 03 16:20:06 crc kubenswrapper[4998]: E1203 16:20:06.342798 4998 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 16:20:06 crc kubenswrapper[4998]: E1203 16:20:06.343444 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/42a2439a-98ee-4172-813c-9ce7109fb1c0-cert podName:42a2439a-98ee-4172-813c-9ce7109fb1c0 nodeName:}" failed. No retries permitted until 2025-12-03 16:20:07.343426581 +0000 UTC m=+985.955126804 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/42a2439a-98ee-4172-813c-9ce7109fb1c0-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" (UID: "42a2439a-98ee-4172-813c-9ce7109fb1c0") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 16:20:06 crc kubenswrapper[4998]: E1203 16:20:06.342899 4998 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 16:20:06 crc kubenswrapper[4998]: E1203 16:20:06.343741 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-metrics-certs podName:07d50ab9-95b8-4c58-9192-445ec4e32b59 nodeName:}" failed. No retries permitted until 2025-12-03 16:20:06.843731778 +0000 UTC m=+985.455432001 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-metrics-certs") pod "openstack-operator-controller-manager-67847ddbd8-t8wbd" (UID: "07d50ab9-95b8-4c58-9192-445ec4e32b59") : secret "metrics-server-cert" not found Dec 03 16:20:06 crc kubenswrapper[4998]: E1203 16:20:06.342904 4998 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 16:20:06 crc kubenswrapper[4998]: E1203 16:20:06.343782 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-webhook-certs podName:07d50ab9-95b8-4c58-9192-445ec4e32b59 nodeName:}" failed. No retries permitted until 2025-12-03 16:20:06.84377635 +0000 UTC m=+985.455476573 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-webhook-certs") pod "openstack-operator-controller-manager-67847ddbd8-t8wbd" (UID: "07d50ab9-95b8-4c58-9192-445ec4e32b59") : secret "webhook-server-cert" not found Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.416775 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjnhc\" (UniqueName: \"kubernetes.io/projected/9f587468-9ae1-43af-8652-f4394865ff37-kube-api-access-vjnhc\") pod \"rabbitmq-cluster-operator-manager-668c99d594-fqdqg\" (UID: \"9f587468-9ae1-43af-8652-f4394865ff37\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fqdqg" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.418526 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szw9p\" (UniqueName: \"kubernetes.io/projected/07d50ab9-95b8-4c58-9192-445ec4e32b59-kube-api-access-szw9p\") pod \"openstack-operator-controller-manager-67847ddbd8-t8wbd\" (UID: \"07d50ab9-95b8-4c58-9192-445ec4e32b59\") " pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.441031 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fqdqg" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.506500 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-mn69l"] Dec 03 16:20:06 crc kubenswrapper[4998]: W1203 16:20:06.562681 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc92de5b9_8ef5_4f4c_b881_82c5f05cd853.slice/crio-7f483bfe407898906907a84ee673fe4a2133de3eb197d5764e408fae91cdd109 WatchSource:0}: Error finding container 7f483bfe407898906907a84ee673fe4a2133de3eb197d5764e408fae91cdd109: Status 404 returned error can't find the container with id 7f483bfe407898906907a84ee673fe4a2133de3eb197d5764e408fae91cdd109 Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.775042 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-tkz2j" event={"ID":"624bac01-c09f-424f-9000-feb138765cc4","Type":"ContainerStarted","Data":"9c0929175a732f66186a750876ee82e511bca68c147c5e16b3bc840b27c7a8da"} Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.776606 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-mn69l" event={"ID":"c92de5b9-8ef5-4f4c-b881-82c5f05cd853","Type":"ContainerStarted","Data":"7f483bfe407898906907a84ee673fe4a2133de3eb197d5764e408fae91cdd109"} Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.783815 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-tcltm" event={"ID":"17cdd58c-7a0c-42e0-bcbe-4545050637e3","Type":"ContainerStarted","Data":"f957039b72d16be052af0220b1c816659ce7330ccb170e68135238eaba1072bb"} Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.852355 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-webhook-certs\") pod \"openstack-operator-controller-manager-67847ddbd8-t8wbd\" (UID: \"07d50ab9-95b8-4c58-9192-445ec4e32b59\") " pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.852412 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-metrics-certs\") pod \"openstack-operator-controller-manager-67847ddbd8-t8wbd\" (UID: \"07d50ab9-95b8-4c58-9192-445ec4e32b59\") " pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" Dec 03 16:20:06 crc kubenswrapper[4998]: E1203 16:20:06.852539 4998 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 16:20:06 crc kubenswrapper[4998]: E1203 16:20:06.852620 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-webhook-certs podName:07d50ab9-95b8-4c58-9192-445ec4e32b59 nodeName:}" failed. No retries permitted until 2025-12-03 16:20:07.852599768 +0000 UTC m=+986.464299981 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-webhook-certs") pod "openstack-operator-controller-manager-67847ddbd8-t8wbd" (UID: "07d50ab9-95b8-4c58-9192-445ec4e32b59") : secret "webhook-server-cert" not found Dec 03 16:20:06 crc kubenswrapper[4998]: E1203 16:20:06.852696 4998 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 16:20:06 crc kubenswrapper[4998]: E1203 16:20:06.852804 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-metrics-certs podName:07d50ab9-95b8-4c58-9192-445ec4e32b59 nodeName:}" failed. No retries permitted until 2025-12-03 16:20:07.852771922 +0000 UTC m=+986.464472235 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-metrics-certs") pod "openstack-operator-controller-manager-67847ddbd8-t8wbd" (UID: "07d50ab9-95b8-4c58-9192-445ec4e32b59") : secret "metrics-server-cert" not found Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.880995 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-m4gc7"] Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.918566 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-tv6bd"] Dec 03 16:20:06 crc kubenswrapper[4998]: W1203 16:20:06.923680 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda67cb778_507e_4dba_9ae4_60ae60315a0a.slice/crio-48fd31194459924b5a70d723dfa851d821c95a71623b2f66f5eef131702a8fb4 WatchSource:0}: Error finding container 48fd31194459924b5a70d723dfa851d821c95a71623b2f66f5eef131702a8fb4: Status 404 returned error can't find the container with id 48fd31194459924b5a70d723dfa851d821c95a71623b2f66f5eef131702a8fb4 Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.934118 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-nfm6p"] Dec 03 16:20:06 crc kubenswrapper[4998]: W1203 16:20:06.945714 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c9761a8_29fb_4d99_9bc0_914a041b36ac.slice/crio-e86e857c5389eb654d5be36c40538f6e8adabc758e78102a7cf71d2c0c9bf2c8 WatchSource:0}: Error finding container e86e857c5389eb654d5be36c40538f6e8adabc758e78102a7cf71d2c0c9bf2c8: Status 404 returned error can't find the container with id e86e857c5389eb654d5be36c40538f6e8adabc758e78102a7cf71d2c0c9bf2c8 Dec 03 16:20:06 crc kubenswrapper[4998]: W1203 16:20:06.945956 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddcb7317b_497a_4135_be25_3462a91406e9.slice/crio-7d2e2a8466dfd5e9a83dfde218dd9a250f1b582edd3349c6eda117bec60268b3 WatchSource:0}: Error finding container 7d2e2a8466dfd5e9a83dfde218dd9a250f1b582edd3349c6eda117bec60268b3: Status 404 returned error can't find the container with id 7d2e2a8466dfd5e9a83dfde218dd9a250f1b582edd3349c6eda117bec60268b3 Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.960939 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f4jqc"] Dec 03 16:20:06 crc kubenswrapper[4998]: I1203 16:20:06.964523 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-5h65w"] Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.057138 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ade59116-1aaf-44e0-b47b-15bdc7b402f0-cert\") pod \"infra-operator-controller-manager-57548d458d-hr74g\" (UID: \"ade59116-1aaf-44e0-b47b-15bdc7b402f0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-hr74g" Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.057283 4998 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.057342 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ade59116-1aaf-44e0-b47b-15bdc7b402f0-cert podName:ade59116-1aaf-44e0-b47b-15bdc7b402f0 nodeName:}" failed. No retries permitted until 2025-12-03 16:20:09.057322761 +0000 UTC m=+987.669022984 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ade59116-1aaf-44e0-b47b-15bdc7b402f0-cert") pod "infra-operator-controller-manager-57548d458d-hr74g" (UID: "ade59116-1aaf-44e0-b47b-15bdc7b402f0") : secret "infra-operator-webhook-server-cert" not found Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.062933 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-f6vdv"] Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.070267 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-h72w5"] Dec 03 16:20:07 crc kubenswrapper[4998]: W1203 16:20:07.071881 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode52d7801_a5c9_47a3_bedc_27174cfbd993.slice/crio-c940fa0a0400082ee5434fe4011d9770489f313d87070360f2bbc690496571c0 WatchSource:0}: Error finding container c940fa0a0400082ee5434fe4011d9770489f313d87070360f2bbc690496571c0: Status 404 returned error can't find the container with id c940fa0a0400082ee5434fe4011d9770489f313d87070360f2bbc690496571c0 Dec 03 16:20:07 crc kubenswrapper[4998]: W1203 16:20:07.075853 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd926336a_658f_49bd_81b9_8086c8bb97a5.slice/crio-0fb95383e2050c9ba41275454c59b12453160068138db2fe4ae3f7d18f20dfb4 WatchSource:0}: Error finding container 0fb95383e2050c9ba41275454c59b12453160068138db2fe4ae3f7d18f20dfb4: Status 404 returned error can't find the container with id 0fb95383e2050c9ba41275454c59b12453160068138db2fe4ae3f7d18f20dfb4 Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.084034 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d2gcr"] Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.090870 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gzg25"] Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.096372 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-ph8r8"] Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.198450 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-z8g4g"] Dec 03 16:20:07 crc kubenswrapper[4998]: W1203 16:20:07.209030 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b89567b_4963_4113_9900_a132d873d260.slice/crio-61778db7761cbbbdfe4df76724719107d187a0a00cbc87f49f2049427aa5cc9a WatchSource:0}: Error finding container 61778db7761cbbbdfe4df76724719107d187a0a00cbc87f49f2049427aa5cc9a: Status 404 returned error can't find the container with id 61778db7761cbbbdfe4df76724719107d187a0a00cbc87f49f2049427aa5cc9a Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.216173 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-2h2m5"] Dec 03 16:20:07 crc kubenswrapper[4998]: W1203 16:20:07.221466 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf23eb2b1_107d_4167_a695_4479df0ae92a.slice/crio-58e35d764ead691f4a439ecd4970c51ec594fd3b7f5f478b556aec93240ce3e7 WatchSource:0}: Error finding container 58e35d764ead691f4a439ecd4970c51ec594fd3b7f5f478b556aec93240ce3e7: Status 404 returned error can't find the container with id 58e35d764ead691f4a439ecd4970c51ec594fd3b7f5f478b556aec93240ce3e7 Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.224232 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wjj8q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-2h2m5_openstack-operators(f23eb2b1-107d-4167-a695-4479df0ae92a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.228988 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-f77f967dc-ndj9c"] Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.231398 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.32:5001/openstack-k8s-operators/watcher-operator:164610071c31587da7ddb7356abc05e8c519dbfa,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xrd78,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-f77f967dc-ndj9c_openstack-operators(ac213ee6-009b-4098-9d18-6dfc8cb97a0e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.231443 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wjj8q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-2h2m5_openstack-operators(f23eb2b1-107d-4167-a695-4479df0ae92a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.232844 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2h2m5" podUID="f23eb2b1-107d-4167-a695-4479df0ae92a" Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.243476 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xrd78,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-f77f967dc-ndj9c_openstack-operators(ac213ee6-009b-4098-9d18-6dfc8cb97a0e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.244888 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-f77f967dc-ndj9c" podUID="ac213ee6-009b-4098-9d18-6dfc8cb97a0e" Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.257570 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fqdqg"] Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.262181 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vjnhc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-fqdqg_openstack-operators(9f587468-9ae1-43af-8652-f4394865ff37): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.263364 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fqdqg" podUID="9f587468-9ae1-43af-8652-f4394865ff37" Dec 03 16:20:07 crc kubenswrapper[4998]: W1203 16:20:07.263913 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e92e441_8150_4817_8039_9cdc8424e562.slice/crio-8de49aba67a7917098aef631bb2747a80915aef8a89c70a1bddc24eb8e4d7dcc WatchSource:0}: Error finding container 8de49aba67a7917098aef631bb2747a80915aef8a89c70a1bddc24eb8e4d7dcc: Status 404 returned error can't find the container with id 8de49aba67a7917098aef631bb2747a80915aef8a89c70a1bddc24eb8e4d7dcc Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.274231 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hzl7f"] Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.275291 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ttf7z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-hzl7f_openstack-operators(0e92e441-8150-4817-8039-9cdc8424e562): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.275525 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pjjsx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-2wrqc_openstack-operators(d7adeced-3036-49a7-8806-fdbefce6406f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.277230 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ttf7z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-hzl7f_openstack-operators(0e92e441-8150-4817-8039-9cdc8424e562): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.278499 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hzl7f" podUID="0e92e441-8150-4817-8039-9cdc8424e562" Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.279515 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lcx75,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-28xct_openstack-operators(79a1881c-77c7-4c5c-b2fc-b2d7d9c31496): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.280383 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2wrqc"] Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.279461 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pjjsx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-2wrqc_openstack-operators(d7adeced-3036-49a7-8806-fdbefce6406f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.281469 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lcx75,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-28xct_openstack-operators(79a1881c-77c7-4c5c-b2fc-b2d7d9c31496): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.282804 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2wrqc" podUID="d7adeced-3036-49a7-8806-fdbefce6406f" Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.282790 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-28xct" podUID="79a1881c-77c7-4c5c-b2fc-b2d7d9c31496" Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.286084 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-28xct"] Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.363449 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42a2439a-98ee-4172-813c-9ce7109fb1c0-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz\" (UID: \"42a2439a-98ee-4172-813c-9ce7109fb1c0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.363646 4998 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.363701 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/42a2439a-98ee-4172-813c-9ce7109fb1c0-cert podName:42a2439a-98ee-4172-813c-9ce7109fb1c0 nodeName:}" failed. No retries permitted until 2025-12-03 16:20:09.363684083 +0000 UTC m=+987.975384306 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/42a2439a-98ee-4172-813c-9ce7109fb1c0-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" (UID: "42a2439a-98ee-4172-813c-9ce7109fb1c0") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.790557 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-28xct" event={"ID":"79a1881c-77c7-4c5c-b2fc-b2d7d9c31496","Type":"ContainerStarted","Data":"a190872718fd253ee5daf01112b813a04f39189dfdb789ef07128a97d09f4876"} Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.793697 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hzl7f" event={"ID":"0e92e441-8150-4817-8039-9cdc8424e562","Type":"ContainerStarted","Data":"8de49aba67a7917098aef631bb2747a80915aef8a89c70a1bddc24eb8e4d7dcc"} Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.795311 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-tv6bd" event={"ID":"a67cb778-507e-4dba-9ae4-60ae60315a0a","Type":"ContainerStarted","Data":"48fd31194459924b5a70d723dfa851d821c95a71623b2f66f5eef131702a8fb4"} Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.796511 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hzl7f" podUID="0e92e441-8150-4817-8039-9cdc8424e562" Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.796607 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nfm6p" event={"ID":"dcb7317b-497a-4135-be25-3462a91406e9","Type":"ContainerStarted","Data":"7d2e2a8466dfd5e9a83dfde218dd9a250f1b582edd3349c6eda117bec60268b3"} Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.799261 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2wrqc" event={"ID":"d7adeced-3036-49a7-8806-fdbefce6406f","Type":"ContainerStarted","Data":"2d147248e0bfc675b83e1350d963e27b023fb782c58cb8df5eec14b4b9cc45e5"} Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.802109 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-28xct" podUID="79a1881c-77c7-4c5c-b2fc-b2d7d9c31496" Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.803459 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2wrqc" podUID="d7adeced-3036-49a7-8806-fdbefce6406f" Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.804372 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2h2m5" event={"ID":"f23eb2b1-107d-4167-a695-4479df0ae92a","Type":"ContainerStarted","Data":"58e35d764ead691f4a439ecd4970c51ec594fd3b7f5f478b556aec93240ce3e7"} Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.806366 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-f6vdv" event={"ID":"aa3ac0c7-ea42-4728-8406-0ce3f9a96c21","Type":"ContainerStarted","Data":"ca8dccfe365b5f1ca11116dc9ebc986b0b054f41d9563e41c35f63ea0de4c1af"} Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.808379 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2h2m5" podUID="f23eb2b1-107d-4167-a695-4479df0ae92a" Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.808608 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fqdqg" event={"ID":"9f587468-9ae1-43af-8652-f4394865ff37","Type":"ContainerStarted","Data":"e235d36e0f0acafeccda39c1641ea347473609d845fb0a0f80ccbe5cdad2e84c"} Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.811690 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fqdqg" podUID="9f587468-9ae1-43af-8652-f4394865ff37" Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.817512 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-f77f967dc-ndj9c" event={"ID":"ac213ee6-009b-4098-9d18-6dfc8cb97a0e","Type":"ContainerStarted","Data":"d9cd136fdab8461751cdb6b4f5bcf8365ee56cb38a9f5ce0ebadd65809ea57f1"} Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.819703 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-ph8r8" event={"ID":"d926336a-658f-49bd-81b9-8086c8bb97a5","Type":"ContainerStarted","Data":"0fb95383e2050c9ba41275454c59b12453160068138db2fe4ae3f7d18f20dfb4"} Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.820643 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.32:5001/openstack-k8s-operators/watcher-operator:164610071c31587da7ddb7356abc05e8c519dbfa\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-f77f967dc-ndj9c" podUID="ac213ee6-009b-4098-9d18-6dfc8cb97a0e" Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.821957 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f4jqc" event={"ID":"0c9761a8-29fb-4d99-9bc0-914a041b36ac","Type":"ContainerStarted","Data":"e86e857c5389eb654d5be36c40538f6e8adabc758e78102a7cf71d2c0c9bf2c8"} Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.823902 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-h72w5" event={"ID":"77fb606c-dea3-4e34-9f4c-5bb305e3b7f4","Type":"ContainerStarted","Data":"95669951e16901d036740f1488f3127a810768359ead86bce8c987cb183fac1f"} Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.828287 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gzg25" event={"ID":"7c372d0c-4cac-4e4c-bc8f-409b5c151d53","Type":"ContainerStarted","Data":"56f83584d6e17d9a9b77df9a41c38a2c99b86cd2c9d30b01adf29a29954ce417"} Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.832063 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5h65w" event={"ID":"da7a11af-1959-4c47-96d3-4c5f4c469493","Type":"ContainerStarted","Data":"d31a242554037c9fda35e028cc368c46ba7f00d08ec6dd5dd9cfcabe6962f412"} Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.834261 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z8g4g" event={"ID":"6b89567b-4963-4113-9900-a132d873d260","Type":"ContainerStarted","Data":"61778db7761cbbbdfe4df76724719107d187a0a00cbc87f49f2049427aa5cc9a"} Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.835017 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-m4gc7" event={"ID":"67e57367-12f9-487d-b691-ca243a28b056","Type":"ContainerStarted","Data":"c5e5916f6250fccf34c9fcf9344de626acbfd03d6c1eb2afcb8d0334cbb40409"} Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.835891 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d2gcr" event={"ID":"e52d7801-a5c9-47a3-bedc-27174cfbd993","Type":"ContainerStarted","Data":"c940fa0a0400082ee5434fe4011d9770489f313d87070360f2bbc690496571c0"} Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.872118 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-webhook-certs\") pod \"openstack-operator-controller-manager-67847ddbd8-t8wbd\" (UID: \"07d50ab9-95b8-4c58-9192-445ec4e32b59\") " pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" Dec 03 16:20:07 crc kubenswrapper[4998]: I1203 16:20:07.872195 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-metrics-certs\") pod \"openstack-operator-controller-manager-67847ddbd8-t8wbd\" (UID: \"07d50ab9-95b8-4c58-9192-445ec4e32b59\") " pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.872422 4998 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.872495 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-metrics-certs podName:07d50ab9-95b8-4c58-9192-445ec4e32b59 nodeName:}" failed. No retries permitted until 2025-12-03 16:20:09.87246241 +0000 UTC m=+988.484162633 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-metrics-certs") pod "openstack-operator-controller-manager-67847ddbd8-t8wbd" (UID: "07d50ab9-95b8-4c58-9192-445ec4e32b59") : secret "metrics-server-cert" not found Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.872547 4998 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 16:20:07 crc kubenswrapper[4998]: E1203 16:20:07.872648 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-webhook-certs podName:07d50ab9-95b8-4c58-9192-445ec4e32b59 nodeName:}" failed. No retries permitted until 2025-12-03 16:20:09.872619714 +0000 UTC m=+988.484319937 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-webhook-certs") pod "openstack-operator-controller-manager-67847ddbd8-t8wbd" (UID: "07d50ab9-95b8-4c58-9192-445ec4e32b59") : secret "webhook-server-cert" not found Dec 03 16:20:08 crc kubenswrapper[4998]: E1203 16:20:08.869076 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fqdqg" podUID="9f587468-9ae1-43af-8652-f4394865ff37" Dec 03 16:20:08 crc kubenswrapper[4998]: E1203 16:20:08.872957 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2h2m5" podUID="f23eb2b1-107d-4167-a695-4479df0ae92a" Dec 03 16:20:08 crc kubenswrapper[4998]: E1203 16:20:08.873111 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.32:5001/openstack-k8s-operators/watcher-operator:164610071c31587da7ddb7356abc05e8c519dbfa\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-f77f967dc-ndj9c" podUID="ac213ee6-009b-4098-9d18-6dfc8cb97a0e" Dec 03 16:20:08 crc kubenswrapper[4998]: E1203 16:20:08.873245 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2wrqc" podUID="d7adeced-3036-49a7-8806-fdbefce6406f" Dec 03 16:20:08 crc kubenswrapper[4998]: E1203 16:20:08.877968 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-28xct" podUID="79a1881c-77c7-4c5c-b2fc-b2d7d9c31496" Dec 03 16:20:08 crc kubenswrapper[4998]: E1203 16:20:08.878074 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hzl7f" podUID="0e92e441-8150-4817-8039-9cdc8424e562" Dec 03 16:20:09 crc kubenswrapper[4998]: I1203 16:20:09.094392 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ade59116-1aaf-44e0-b47b-15bdc7b402f0-cert\") pod \"infra-operator-controller-manager-57548d458d-hr74g\" (UID: \"ade59116-1aaf-44e0-b47b-15bdc7b402f0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-hr74g" Dec 03 16:20:09 crc kubenswrapper[4998]: E1203 16:20:09.094625 4998 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 16:20:09 crc kubenswrapper[4998]: E1203 16:20:09.094675 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ade59116-1aaf-44e0-b47b-15bdc7b402f0-cert podName:ade59116-1aaf-44e0-b47b-15bdc7b402f0 nodeName:}" failed. No retries permitted until 2025-12-03 16:20:13.094661366 +0000 UTC m=+991.706361589 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ade59116-1aaf-44e0-b47b-15bdc7b402f0-cert") pod "infra-operator-controller-manager-57548d458d-hr74g" (UID: "ade59116-1aaf-44e0-b47b-15bdc7b402f0") : secret "infra-operator-webhook-server-cert" not found Dec 03 16:20:09 crc kubenswrapper[4998]: I1203 16:20:09.399204 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42a2439a-98ee-4172-813c-9ce7109fb1c0-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz\" (UID: \"42a2439a-98ee-4172-813c-9ce7109fb1c0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" Dec 03 16:20:09 crc kubenswrapper[4998]: E1203 16:20:09.399347 4998 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 16:20:09 crc kubenswrapper[4998]: E1203 16:20:09.399418 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/42a2439a-98ee-4172-813c-9ce7109fb1c0-cert podName:42a2439a-98ee-4172-813c-9ce7109fb1c0 nodeName:}" failed. No retries permitted until 2025-12-03 16:20:13.399398506 +0000 UTC m=+992.011098729 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/42a2439a-98ee-4172-813c-9ce7109fb1c0-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" (UID: "42a2439a-98ee-4172-813c-9ce7109fb1c0") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 16:20:09 crc kubenswrapper[4998]: I1203 16:20:09.906947 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-webhook-certs\") pod \"openstack-operator-controller-manager-67847ddbd8-t8wbd\" (UID: \"07d50ab9-95b8-4c58-9192-445ec4e32b59\") " pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" Dec 03 16:20:09 crc kubenswrapper[4998]: I1203 16:20:09.907013 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-metrics-certs\") pod \"openstack-operator-controller-manager-67847ddbd8-t8wbd\" (UID: \"07d50ab9-95b8-4c58-9192-445ec4e32b59\") " pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" Dec 03 16:20:09 crc kubenswrapper[4998]: E1203 16:20:09.907096 4998 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 16:20:09 crc kubenswrapper[4998]: E1203 16:20:09.907138 4998 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 16:20:09 crc kubenswrapper[4998]: E1203 16:20:09.907178 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-webhook-certs podName:07d50ab9-95b8-4c58-9192-445ec4e32b59 nodeName:}" failed. No retries permitted until 2025-12-03 16:20:13.907160087 +0000 UTC m=+992.518860310 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-webhook-certs") pod "openstack-operator-controller-manager-67847ddbd8-t8wbd" (UID: "07d50ab9-95b8-4c58-9192-445ec4e32b59") : secret "webhook-server-cert" not found Dec 03 16:20:09 crc kubenswrapper[4998]: E1203 16:20:09.907212 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-metrics-certs podName:07d50ab9-95b8-4c58-9192-445ec4e32b59 nodeName:}" failed. No retries permitted until 2025-12-03 16:20:13.907187798 +0000 UTC m=+992.518888021 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-metrics-certs") pod "openstack-operator-controller-manager-67847ddbd8-t8wbd" (UID: "07d50ab9-95b8-4c58-9192-445ec4e32b59") : secret "metrics-server-cert" not found Dec 03 16:20:13 crc kubenswrapper[4998]: I1203 16:20:13.158882 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ade59116-1aaf-44e0-b47b-15bdc7b402f0-cert\") pod \"infra-operator-controller-manager-57548d458d-hr74g\" (UID: \"ade59116-1aaf-44e0-b47b-15bdc7b402f0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-hr74g" Dec 03 16:20:13 crc kubenswrapper[4998]: E1203 16:20:13.159334 4998 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 16:20:13 crc kubenswrapper[4998]: E1203 16:20:13.159830 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ade59116-1aaf-44e0-b47b-15bdc7b402f0-cert podName:ade59116-1aaf-44e0-b47b-15bdc7b402f0 nodeName:}" failed. No retries permitted until 2025-12-03 16:20:21.159804842 +0000 UTC m=+999.771505065 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ade59116-1aaf-44e0-b47b-15bdc7b402f0-cert") pod "infra-operator-controller-manager-57548d458d-hr74g" (UID: "ade59116-1aaf-44e0-b47b-15bdc7b402f0") : secret "infra-operator-webhook-server-cert" not found Dec 03 16:20:13 crc kubenswrapper[4998]: I1203 16:20:13.463916 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42a2439a-98ee-4172-813c-9ce7109fb1c0-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz\" (UID: \"42a2439a-98ee-4172-813c-9ce7109fb1c0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" Dec 03 16:20:13 crc kubenswrapper[4998]: E1203 16:20:13.464141 4998 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 16:20:13 crc kubenswrapper[4998]: E1203 16:20:13.464275 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/42a2439a-98ee-4172-813c-9ce7109fb1c0-cert podName:42a2439a-98ee-4172-813c-9ce7109fb1c0 nodeName:}" failed. No retries permitted until 2025-12-03 16:20:21.464241686 +0000 UTC m=+1000.075941979 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/42a2439a-98ee-4172-813c-9ce7109fb1c0-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" (UID: "42a2439a-98ee-4172-813c-9ce7109fb1c0") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 16:20:13 crc kubenswrapper[4998]: I1203 16:20:13.972244 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-webhook-certs\") pod \"openstack-operator-controller-manager-67847ddbd8-t8wbd\" (UID: \"07d50ab9-95b8-4c58-9192-445ec4e32b59\") " pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" Dec 03 16:20:13 crc kubenswrapper[4998]: I1203 16:20:13.972312 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-metrics-certs\") pod \"openstack-operator-controller-manager-67847ddbd8-t8wbd\" (UID: \"07d50ab9-95b8-4c58-9192-445ec4e32b59\") " pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" Dec 03 16:20:13 crc kubenswrapper[4998]: E1203 16:20:13.972466 4998 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 16:20:13 crc kubenswrapper[4998]: E1203 16:20:13.972491 4998 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 16:20:13 crc kubenswrapper[4998]: E1203 16:20:13.972555 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-webhook-certs podName:07d50ab9-95b8-4c58-9192-445ec4e32b59 nodeName:}" failed. No retries permitted until 2025-12-03 16:20:21.97253375 +0000 UTC m=+1000.584233963 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-webhook-certs") pod "openstack-operator-controller-manager-67847ddbd8-t8wbd" (UID: "07d50ab9-95b8-4c58-9192-445ec4e32b59") : secret "webhook-server-cert" not found Dec 03 16:20:13 crc kubenswrapper[4998]: E1203 16:20:13.972574 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-metrics-certs podName:07d50ab9-95b8-4c58-9192-445ec4e32b59 nodeName:}" failed. No retries permitted until 2025-12-03 16:20:21.972566411 +0000 UTC m=+1000.584266634 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-metrics-certs") pod "openstack-operator-controller-manager-67847ddbd8-t8wbd" (UID: "07d50ab9-95b8-4c58-9192-445ec4e32b59") : secret "metrics-server-cert" not found Dec 03 16:20:20 crc kubenswrapper[4998]: E1203 16:20:20.347815 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94" Dec 03 16:20:20 crc kubenswrapper[4998]: E1203 16:20:20.348484 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gkd5c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-z8g4g_openstack-operators(6b89567b-4963-4113-9900-a132d873d260): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 16:20:20 crc kubenswrapper[4998]: E1203 16:20:20.966296 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fjrsz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-5h65w_openstack-operators(da7a11af-1959-4c47-96d3-4c5f4c469493): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 16:20:20 crc kubenswrapper[4998]: E1203 16:20:20.967709 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5h65w" podUID="da7a11af-1959-4c47-96d3-4c5f4c469493" Dec 03 16:20:20 crc kubenswrapper[4998]: I1203 16:20:20.972809 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-mn69l" event={"ID":"c92de5b9-8ef5-4f4c-b881-82c5f05cd853","Type":"ContainerStarted","Data":"255bcec352642f163f47af1fa988aa0ed295895de2a7ea2cb4e99e7b7a0b227b"} Dec 03 16:20:20 crc kubenswrapper[4998]: I1203 16:20:20.976865 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5h65w" event={"ID":"da7a11af-1959-4c47-96d3-4c5f4c469493","Type":"ContainerStarted","Data":"ca541e9e1c0c36818b447a7515e10a7844333747b38c97c1d2aae2aaef444fee"} Dec 03 16:20:20 crc kubenswrapper[4998]: I1203 16:20:20.976951 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5h65w" Dec 03 16:20:20 crc kubenswrapper[4998]: E1203 16:20:20.979951 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5h65w" podUID="da7a11af-1959-4c47-96d3-4c5f4c469493" Dec 03 16:20:20 crc kubenswrapper[4998]: I1203 16:20:20.985625 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-ph8r8" event={"ID":"d926336a-658f-49bd-81b9-8086c8bb97a5","Type":"ContainerStarted","Data":"6494bd103bdb36d8e79fef3e7358ec8665ccfe7d43ff06a20d30480d501bc06a"} Dec 03 16:20:20 crc kubenswrapper[4998]: E1203 16:20:20.987067 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vmdr6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-f4jqc_openstack-operators(0c9761a8-29fb-4d99-9bc0-914a041b36ac): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 16:20:20 crc kubenswrapper[4998]: E1203 16:20:20.988573 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f4jqc" podUID="0c9761a8-29fb-4d99-9bc0-914a041b36ac" Dec 03 16:20:21 crc kubenswrapper[4998]: I1203 16:20:21.016112 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-h72w5" event={"ID":"77fb606c-dea3-4e34-9f4c-5bb305e3b7f4","Type":"ContainerStarted","Data":"4fd9e72f7770ce926be8b34cf0c64c59e4892b5ee960276e730676f6f9f03f3d"} Dec 03 16:20:21 crc kubenswrapper[4998]: I1203 16:20:21.026968 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-tcltm" event={"ID":"17cdd58c-7a0c-42e0-bcbe-4545050637e3","Type":"ContainerStarted","Data":"f5cf6558b90679050d6c86201e0b05c02aadf30469f644482e7d134a2a58717d"} Dec 03 16:20:21 crc kubenswrapper[4998]: I1203 16:20:21.034520 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-f6vdv" event={"ID":"aa3ac0c7-ea42-4728-8406-0ce3f9a96c21","Type":"ContainerStarted","Data":"009d849b2c262c96056dbd0e1dd5d86c55b9a1efe152fd136c1952119d7b44c7"} Dec 03 16:20:21 crc kubenswrapper[4998]: I1203 16:20:21.039672 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gzg25" event={"ID":"7c372d0c-4cac-4e4c-bc8f-409b5c151d53","Type":"ContainerStarted","Data":"a59b68cc4d177eb7a65f4a8a4cd9af20d36ba88a7978fe862d8722f62547bf67"} Dec 03 16:20:21 crc kubenswrapper[4998]: I1203 16:20:21.041992 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-tv6bd" event={"ID":"a67cb778-507e-4dba-9ae4-60ae60315a0a","Type":"ContainerStarted","Data":"327a097f7a8f211ba5d3e638fc56c15a5110bb4055dd364257df92296f98cd1c"} Dec 03 16:20:21 crc kubenswrapper[4998]: E1203 16:20:21.043198 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qhjfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-nfm6p_openstack-operators(dcb7317b-497a-4135-be25-3462a91406e9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 16:20:21 crc kubenswrapper[4998]: E1203 16:20:21.051068 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nfm6p" podUID="dcb7317b-497a-4135-be25-3462a91406e9" Dec 03 16:20:21 crc kubenswrapper[4998]: I1203 16:20:21.051601 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-tkz2j" event={"ID":"624bac01-c09f-424f-9000-feb138765cc4","Type":"ContainerStarted","Data":"956b9cf6819efc2d7057c1c6b85614a04d9b1f9a1d42e0745afbb80a238dec15"} Dec 03 16:20:21 crc kubenswrapper[4998]: I1203 16:20:21.058059 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f4jqc" Dec 03 16:20:21 crc kubenswrapper[4998]: I1203 16:20:21.069277 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-m4gc7" event={"ID":"67e57367-12f9-487d-b691-ca243a28b056","Type":"ContainerStarted","Data":"a542d4b6f653e8fc0556f59f91adda4bd18fe12e8817229bb2ca805f153c91bb"} Dec 03 16:20:21 crc kubenswrapper[4998]: E1203 16:20:21.070204 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f4jqc" podUID="0c9761a8-29fb-4d99-9bc0-914a041b36ac" Dec 03 16:20:21 crc kubenswrapper[4998]: I1203 16:20:21.073774 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d2gcr" event={"ID":"e52d7801-a5c9-47a3-bedc-27174cfbd993","Type":"ContainerStarted","Data":"e1d22ef9d8d38e58412f0dc360f394d8c16a25f7ed2b349acbb6a8d46852b473"} Dec 03 16:20:21 crc kubenswrapper[4998]: I1203 16:20:21.188102 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ade59116-1aaf-44e0-b47b-15bdc7b402f0-cert\") pod \"infra-operator-controller-manager-57548d458d-hr74g\" (UID: \"ade59116-1aaf-44e0-b47b-15bdc7b402f0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-hr74g" Dec 03 16:20:21 crc kubenswrapper[4998]: I1203 16:20:21.197285 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ade59116-1aaf-44e0-b47b-15bdc7b402f0-cert\") pod \"infra-operator-controller-manager-57548d458d-hr74g\" (UID: \"ade59116-1aaf-44e0-b47b-15bdc7b402f0\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-hr74g" Dec 03 16:20:21 crc kubenswrapper[4998]: I1203 16:20:21.208059 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-hr74g" Dec 03 16:20:21 crc kubenswrapper[4998]: I1203 16:20:21.492346 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42a2439a-98ee-4172-813c-9ce7109fb1c0-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz\" (UID: \"42a2439a-98ee-4172-813c-9ce7109fb1c0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" Dec 03 16:20:21 crc kubenswrapper[4998]: I1203 16:20:21.499537 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/42a2439a-98ee-4172-813c-9ce7109fb1c0-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz\" (UID: \"42a2439a-98ee-4172-813c-9ce7109fb1c0\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" Dec 03 16:20:21 crc kubenswrapper[4998]: I1203 16:20:21.645636 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" Dec 03 16:20:22 crc kubenswrapper[4998]: I1203 16:20:22.010559 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-webhook-certs\") pod \"openstack-operator-controller-manager-67847ddbd8-t8wbd\" (UID: \"07d50ab9-95b8-4c58-9192-445ec4e32b59\") " pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" Dec 03 16:20:22 crc kubenswrapper[4998]: I1203 16:20:22.010953 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-metrics-certs\") pod \"openstack-operator-controller-manager-67847ddbd8-t8wbd\" (UID: \"07d50ab9-95b8-4c58-9192-445ec4e32b59\") " pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" Dec 03 16:20:22 crc kubenswrapper[4998]: E1203 16:20:22.010746 4998 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 16:20:22 crc kubenswrapper[4998]: E1203 16:20:22.011052 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-webhook-certs podName:07d50ab9-95b8-4c58-9192-445ec4e32b59 nodeName:}" failed. No retries permitted until 2025-12-03 16:20:38.011033866 +0000 UTC m=+1016.622734079 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-webhook-certs") pod "openstack-operator-controller-manager-67847ddbd8-t8wbd" (UID: "07d50ab9-95b8-4c58-9192-445ec4e32b59") : secret "webhook-server-cert" not found Dec 03 16:20:22 crc kubenswrapper[4998]: I1203 16:20:22.016952 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-metrics-certs\") pod \"openstack-operator-controller-manager-67847ddbd8-t8wbd\" (UID: \"07d50ab9-95b8-4c58-9192-445ec4e32b59\") " pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" Dec 03 16:20:22 crc kubenswrapper[4998]: I1203 16:20:22.146463 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f4jqc" event={"ID":"0c9761a8-29fb-4d99-9bc0-914a041b36ac","Type":"ContainerStarted","Data":"b8aafdd411e34ad0f392df765277c97918550a1afbfe24c6021987cc6fad7d67"} Dec 03 16:20:22 crc kubenswrapper[4998]: E1203 16:20:22.149442 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f4jqc" podUID="0c9761a8-29fb-4d99-9bc0-914a041b36ac" Dec 03 16:20:22 crc kubenswrapper[4998]: I1203 16:20:22.153690 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nfm6p" event={"ID":"dcb7317b-497a-4135-be25-3462a91406e9","Type":"ContainerStarted","Data":"c744602be565213c1e568835538857e83ea4f539efa84ce13ee19aedb7229c85"} Dec 03 16:20:22 crc kubenswrapper[4998]: I1203 16:20:22.153720 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nfm6p" Dec 03 16:20:22 crc kubenswrapper[4998]: E1203 16:20:22.154234 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nfm6p" podUID="dcb7317b-497a-4135-be25-3462a91406e9" Dec 03 16:20:22 crc kubenswrapper[4998]: E1203 16:20:22.154415 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5h65w" podUID="da7a11af-1959-4c47-96d3-4c5f4c469493" Dec 03 16:20:22 crc kubenswrapper[4998]: I1203 16:20:22.332313 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-hr74g"] Dec 03 16:20:22 crc kubenswrapper[4998]: I1203 16:20:22.378366 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz"] Dec 03 16:20:22 crc kubenswrapper[4998]: W1203 16:20:22.486117 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podade59116_1aaf_44e0_b47b_15bdc7b402f0.slice/crio-4239076803aeb33ff0da8f64b68ef1311e87f74fc0826d1eac3eacb1ae7b74fe WatchSource:0}: Error finding container 4239076803aeb33ff0da8f64b68ef1311e87f74fc0826d1eac3eacb1ae7b74fe: Status 404 returned error can't find the container with id 4239076803aeb33ff0da8f64b68ef1311e87f74fc0826d1eac3eacb1ae7b74fe Dec 03 16:20:23 crc kubenswrapper[4998]: I1203 16:20:23.159356 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-hr74g" event={"ID":"ade59116-1aaf-44e0-b47b-15bdc7b402f0","Type":"ContainerStarted","Data":"4239076803aeb33ff0da8f64b68ef1311e87f74fc0826d1eac3eacb1ae7b74fe"} Dec 03 16:20:23 crc kubenswrapper[4998]: I1203 16:20:23.161515 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" event={"ID":"42a2439a-98ee-4172-813c-9ce7109fb1c0","Type":"ContainerStarted","Data":"9e0afc1d13d6aa097b92eb8997ec52cd3e07693e1b076860cb8c95d4c3216243"} Dec 03 16:20:23 crc kubenswrapper[4998]: E1203 16:20:23.163526 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f4jqc" podUID="0c9761a8-29fb-4d99-9bc0-914a041b36ac" Dec 03 16:20:23 crc kubenswrapper[4998]: E1203 16:20:23.164264 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nfm6p" podUID="dcb7317b-497a-4135-be25-3462a91406e9" Dec 03 16:20:25 crc kubenswrapper[4998]: I1203 16:20:25.735624 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5h65w" Dec 03 16:20:25 crc kubenswrapper[4998]: E1203 16:20:25.743322 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5h65w" podUID="da7a11af-1959-4c47-96d3-4c5f4c469493" Dec 03 16:20:25 crc kubenswrapper[4998]: I1203 16:20:25.912568 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f4jqc" Dec 03 16:20:25 crc kubenswrapper[4998]: E1203 16:20:25.916845 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f4jqc" podUID="0c9761a8-29fb-4d99-9bc0-914a041b36ac" Dec 03 16:20:35 crc kubenswrapper[4998]: I1203 16:20:35.735293 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nfm6p" Dec 03 16:20:38 crc kubenswrapper[4998]: I1203 16:20:38.110055 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-webhook-certs\") pod \"openstack-operator-controller-manager-67847ddbd8-t8wbd\" (UID: \"07d50ab9-95b8-4c58-9192-445ec4e32b59\") " pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" Dec 03 16:20:38 crc kubenswrapper[4998]: I1203 16:20:38.119734 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/07d50ab9-95b8-4c58-9192-445ec4e32b59-webhook-certs\") pod \"openstack-operator-controller-manager-67847ddbd8-t8wbd\" (UID: \"07d50ab9-95b8-4c58-9192-445ec4e32b59\") " pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" Dec 03 16:20:38 crc kubenswrapper[4998]: I1203 16:20:38.176628 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" Dec 03 16:20:49 crc kubenswrapper[4998]: E1203 16:20:49.894590 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557" Dec 03 16:20:49 crc kubenswrapper[4998]: E1203 16:20:49.895214 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ttf7z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-hzl7f_openstack-operators(0e92e441-8150-4817-8039-9cdc8424e562): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 16:20:50 crc kubenswrapper[4998]: E1203 16:20:50.350871 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d" Dec 03 16:20:50 crc kubenswrapper[4998]: E1203 16:20:50.351046 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pjjsx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-2wrqc_openstack-operators(d7adeced-3036-49a7-8806-fdbefce6406f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 16:20:50 crc kubenswrapper[4998]: E1203 16:20:50.723263 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7" Dec 03 16:20:50 crc kubenswrapper[4998]: E1203 16:20:50.723488 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cgvv9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-57548d458d-hr74g_openstack-operators(ade59116-1aaf-44e0-b47b-15bdc7b402f0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 16:20:51 crc kubenswrapper[4998]: E1203 16:20:51.143141 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81" Dec 03 16:20:51 crc kubenswrapper[4998]: E1203 16:20:51.143687 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7qvr2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz_openstack-operators(42a2439a-98ee-4172-813c-9ce7109fb1c0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 16:20:51 crc kubenswrapper[4998]: E1203 16:20:51.211739 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.32:5001/openstack-k8s-operators/watcher-operator:164610071c31587da7ddb7356abc05e8c519dbfa" Dec 03 16:20:51 crc kubenswrapper[4998]: E1203 16:20:51.211804 4998 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.32:5001/openstack-k8s-operators/watcher-operator:164610071c31587da7ddb7356abc05e8c519dbfa" Dec 03 16:20:51 crc kubenswrapper[4998]: E1203 16:20:51.211944 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.32:5001/openstack-k8s-operators/watcher-operator:164610071c31587da7ddb7356abc05e8c519dbfa,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xrd78,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-f77f967dc-ndj9c_openstack-operators(ac213ee6-009b-4098-9d18-6dfc8cb97a0e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 16:20:51 crc kubenswrapper[4998]: E1203 16:20:51.647485 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 03 16:20:51 crc kubenswrapper[4998]: E1203 16:20:51.647668 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vjnhc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-fqdqg_openstack-operators(9f587468-9ae1-43af-8652-f4394865ff37): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 16:20:51 crc kubenswrapper[4998]: E1203 16:20:51.648868 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fqdqg" podUID="9f587468-9ae1-43af-8652-f4394865ff37" Dec 03 16:20:51 crc kubenswrapper[4998]: E1203 16:20:51.778997 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 16:20:51 crc kubenswrapper[4998]: E1203 16:20:51.779457 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rrmlx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-tcltm_openstack-operators(17cdd58c-7a0c-42e0-bcbe-4545050637e3): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 03 16:20:51 crc kubenswrapper[4998]: E1203 16:20:51.780312 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 16:20:51 crc kubenswrapper[4998]: E1203 16:20:51.780448 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rsv2h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-h72w5_openstack-operators(77fb606c-dea3-4e34-9f4c-5bb305e3b7f4): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 03 16:20:51 crc kubenswrapper[4998]: E1203 16:20:51.780541 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\": context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 16:20:51 crc kubenswrapper[4998]: E1203 16:20:51.780615 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8nkng,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-d2gcr_openstack-operators(e52d7801-a5c9-47a3-bedc-27174cfbd993): ErrImagePull: rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\": context canceled" logger="UnhandledError" Dec 03 16:20:51 crc kubenswrapper[4998]: E1203 16:20:51.780699 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-tcltm" podUID="17cdd58c-7a0c-42e0-bcbe-4545050637e3" Dec 03 16:20:51 crc kubenswrapper[4998]: E1203 16:20:51.781769 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \\\"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\\\": context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d2gcr" podUID="e52d7801-a5c9-47a3-bedc-27174cfbd993" Dec 03 16:20:51 crc kubenswrapper[4998]: E1203 16:20:51.781807 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-h72w5" podUID="77fb606c-dea3-4e34-9f4c-5bb305e3b7f4" Dec 03 16:20:52 crc kubenswrapper[4998]: I1203 16:20:52.025918 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd"] Dec 03 16:20:52 crc kubenswrapper[4998]: E1203 16:20:52.360723 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 16:20:52 crc kubenswrapper[4998]: E1203 16:20:52.360972 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zlllz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-tkz2j_openstack-operators(624bac01-c09f-424f-9000-feb138765cc4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 16:20:52 crc kubenswrapper[4998]: E1203 16:20:52.362151 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-tkz2j" podUID="624bac01-c09f-424f-9000-feb138765cc4" Dec 03 16:20:52 crc kubenswrapper[4998]: W1203 16:20:52.371242 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07d50ab9_95b8_4c58_9192_445ec4e32b59.slice/crio-5b849cdb558984f2e8825a19ed0ed11fe9fc64390e6bec74c9e9f9dc7de69396 WatchSource:0}: Error finding container 5b849cdb558984f2e8825a19ed0ed11fe9fc64390e6bec74c9e9f9dc7de69396: Status 404 returned error can't find the container with id 5b849cdb558984f2e8825a19ed0ed11fe9fc64390e6bec74c9e9f9dc7de69396 Dec 03 16:20:52 crc kubenswrapper[4998]: I1203 16:20:52.398486 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" event={"ID":"07d50ab9-95b8-4c58-9192-445ec4e32b59","Type":"ContainerStarted","Data":"5b849cdb558984f2e8825a19ed0ed11fe9fc64390e6bec74c9e9f9dc7de69396"} Dec 03 16:20:52 crc kubenswrapper[4998]: I1203 16:20:52.398781 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-h72w5" Dec 03 16:20:52 crc kubenswrapper[4998]: I1203 16:20:52.398823 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-tkz2j" Dec 03 16:20:52 crc kubenswrapper[4998]: I1203 16:20:52.401494 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-tkz2j" Dec 03 16:20:52 crc kubenswrapper[4998]: I1203 16:20:52.402212 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-h72w5" Dec 03 16:20:52 crc kubenswrapper[4998]: E1203 16:20:52.407233 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d2gcr" podUID="e52d7801-a5c9-47a3-bedc-27174cfbd993" Dec 03 16:20:52 crc kubenswrapper[4998]: E1203 16:20:52.407257 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-tcltm" podUID="17cdd58c-7a0c-42e0-bcbe-4545050637e3" Dec 03 16:20:52 crc kubenswrapper[4998]: E1203 16:20:52.407369 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-tkz2j" podUID="624bac01-c09f-424f-9000-feb138765cc4" Dec 03 16:20:52 crc kubenswrapper[4998]: E1203 16:20:52.407397 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-h72w5" podUID="77fb606c-dea3-4e34-9f4c-5bb305e3b7f4" Dec 03 16:20:52 crc kubenswrapper[4998]: E1203 16:20:52.564037 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 16:20:52 crc kubenswrapper[4998]: E1203 16:20:52.564525 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ksr2r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-m4gc7_openstack-operators(67e57367-12f9-487d-b691-ca243a28b056): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 16:20:52 crc kubenswrapper[4998]: E1203 16:20:52.565726 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-m4gc7" podUID="67e57367-12f9-487d-b691-ca243a28b056" Dec 03 16:20:52 crc kubenswrapper[4998]: E1203 16:20:52.693781 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 16:20:52 crc kubenswrapper[4998]: E1203 16:20:52.693969 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-plmts,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-gzg25_openstack-operators(7c372d0c-4cac-4e4c-bc8f-409b5c151d53): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 16:20:52 crc kubenswrapper[4998]: E1203 16:20:52.694821 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 16:20:52 crc kubenswrapper[4998]: E1203 16:20:52.694994 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gkd5c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-z8g4g_openstack-operators(6b89567b-4963-4113-9900-a132d873d260): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 16:20:52 crc kubenswrapper[4998]: E1203 16:20:52.696118 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z8g4g" podUID="6b89567b-4963-4113-9900-a132d873d260" Dec 03 16:20:52 crc kubenswrapper[4998]: E1203 16:20:52.697800 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gzg25" podUID="7c372d0c-4cac-4e4c-bc8f-409b5c151d53" Dec 03 16:20:52 crc kubenswrapper[4998]: E1203 16:20:52.707975 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 16:20:52 crc kubenswrapper[4998]: E1203 16:20:52.708085 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-87l7h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-859b6ccc6-tv6bd_openstack-operators(a67cb778-507e-4dba-9ae4-60ae60315a0a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 16:20:52 crc kubenswrapper[4998]: E1203 16:20:52.709218 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-tv6bd" podUID="a67cb778-507e-4dba-9ae4-60ae60315a0a" Dec 03 16:20:53 crc kubenswrapper[4998]: E1203 16:20:53.004582 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 16:20:53 crc kubenswrapper[4998]: E1203 16:20:53.004746 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7k8kc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7c79b5df47-mn69l_openstack-operators(c92de5b9-8ef5-4f4c-b881-82c5f05cd853): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 16:20:53 crc kubenswrapper[4998]: E1203 16:20:53.005936 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-mn69l" podUID="c92de5b9-8ef5-4f4c-b881-82c5f05cd853" Dec 03 16:20:53 crc kubenswrapper[4998]: E1203 16:20:53.007427 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 16:20:53 crc kubenswrapper[4998]: E1203 16:20:53.007549 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6gcmg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-f6vdv_openstack-operators(aa3ac0c7-ea42-4728-8406-0ce3f9a96c21): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 16:20:53 crc kubenswrapper[4998]: E1203 16:20:53.009249 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-f6vdv" podUID="aa3ac0c7-ea42-4728-8406-0ce3f9a96c21" Dec 03 16:20:53 crc kubenswrapper[4998]: E1203 16:20:53.189329 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 16:20:53 crc kubenswrapper[4998]: E1203 16:20:53.189819 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dwwrh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-ph8r8_openstack-operators(d926336a-658f-49bd-81b9-8086c8bb97a5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 16:20:53 crc kubenswrapper[4998]: E1203 16:20:53.191029 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-ph8r8" podUID="d926336a-658f-49bd-81b9-8086c8bb97a5" Dec 03 16:20:53 crc kubenswrapper[4998]: E1203 16:20:53.233958 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-hr74g" podUID="ade59116-1aaf-44e0-b47b-15bdc7b402f0" Dec 03 16:20:53 crc kubenswrapper[4998]: E1203 16:20:53.244573 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" podUID="42a2439a-98ee-4172-813c-9ce7109fb1c0" Dec 03 16:20:53 crc kubenswrapper[4998]: E1203 16:20:53.252656 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-f77f967dc-ndj9c" podUID="ac213ee6-009b-4098-9d18-6dfc8cb97a0e" Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.431057 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f4jqc" event={"ID":"0c9761a8-29fb-4d99-9bc0-914a041b36ac","Type":"ContainerStarted","Data":"cb4d30464d1d0331af1c1277af478fbe3414664b12b05c3d355572b314e19d09"} Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.448534 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2h2m5" event={"ID":"f23eb2b1-107d-4167-a695-4479df0ae92a","Type":"ContainerStarted","Data":"62d902b0a729d6a6b31af8836f17359039a8c427f069694ade64f1b94b3ddc89"} Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.451391 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nfm6p" event={"ID":"dcb7317b-497a-4135-be25-3462a91406e9","Type":"ContainerStarted","Data":"d5c6283e3c6810112ffee0210c22c25d189d973284a9418fe5072b8ba62218d0"} Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.460572 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-f4jqc" podStartSLOduration=2.761893106 podStartE2EDuration="48.460549673s" podCreationTimestamp="2025-12-03 16:20:05 +0000 UTC" firstStartedPulling="2025-12-03 16:20:06.962019943 +0000 UTC m=+985.573720166" lastFinishedPulling="2025-12-03 16:20:52.66067651 +0000 UTC m=+1031.272376733" observedRunningTime="2025-12-03 16:20:53.455918426 +0000 UTC m=+1032.067618659" watchObservedRunningTime="2025-12-03 16:20:53.460549673 +0000 UTC m=+1032.072249896" Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.464511 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-f77f967dc-ndj9c" event={"ID":"ac213ee6-009b-4098-9d18-6dfc8cb97a0e","Type":"ContainerStarted","Data":"b8be74a4255541354eac9a041a107ddb4e4f2488f9408f0c7ce9413f55f17bc4"} Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.467331 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-28xct" event={"ID":"79a1881c-77c7-4c5c-b2fc-b2d7d9c31496","Type":"ContainerStarted","Data":"00678b768c9fe609d39718d9e49939c8ca6c822bcca9144951b5c56467f9abec"} Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.469654 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-hr74g" event={"ID":"ade59116-1aaf-44e0-b47b-15bdc7b402f0","Type":"ContainerStarted","Data":"5b99cc942c3d7b6fd7343d2c753ae2ab3d2e59d5713a7e75d86e7e8887770296"} Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.472055 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" event={"ID":"42a2439a-98ee-4172-813c-9ce7109fb1c0","Type":"ContainerStarted","Data":"3d495ec54384f08508bd21b8dafe8a47027e5c560a26ffc6e350ee0d382e6a37"} Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.474077 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5h65w" event={"ID":"da7a11af-1959-4c47-96d3-4c5f4c469493","Type":"ContainerStarted","Data":"c12aa6b71c58bc4f9186db2ad8b25451d59d72cae847da2d6acbb3f8645fc0de"} Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.476473 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" event={"ID":"07d50ab9-95b8-4c58-9192-445ec4e32b59","Type":"ContainerStarted","Data":"92ae4285e7b2728a9787ebcb99dae1b7a69636ac280a5727e8e9a19b0c9ce440"} Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.476503 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.478794 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-tv6bd" Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.478827 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-f6vdv" Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.478837 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gzg25" Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.478849 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-m4gc7" Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.479714 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-ph8r8" Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.479881 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-mn69l" Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.481914 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-tv6bd" Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.482105 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gzg25" Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.484927 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-nfm6p" podStartSLOduration=2.79091865 podStartE2EDuration="48.484913629s" podCreationTimestamp="2025-12-03 16:20:05 +0000 UTC" firstStartedPulling="2025-12-03 16:20:06.967225895 +0000 UTC m=+985.578926118" lastFinishedPulling="2025-12-03 16:20:52.661220864 +0000 UTC m=+1031.272921097" observedRunningTime="2025-12-03 16:20:53.483486153 +0000 UTC m=+1032.095186386" watchObservedRunningTime="2025-12-03 16:20:53.484913629 +0000 UTC m=+1032.096613862" Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.498109 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-f6vdv" Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.498386 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-ph8r8" Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.498567 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-mn69l" Dec 03 16:20:53 crc kubenswrapper[4998]: E1203 16:20:53.498670 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7\\\"\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-hr74g" podUID="ade59116-1aaf-44e0-b47b-15bdc7b402f0" Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.499016 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-m4gc7" Dec 03 16:20:53 crc kubenswrapper[4998]: E1203 16:20:53.499194 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" podUID="42a2439a-98ee-4172-813c-9ce7109fb1c0" Dec 03 16:20:53 crc kubenswrapper[4998]: E1203 16:20:53.499249 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.32:5001/openstack-k8s-operators/watcher-operator:164610071c31587da7ddb7356abc05e8c519dbfa\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-f77f967dc-ndj9c" podUID="ac213ee6-009b-4098-9d18-6dfc8cb97a0e" Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.598657 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" podStartSLOduration=48.598639103 podStartE2EDuration="48.598639103s" podCreationTimestamp="2025-12-03 16:20:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:20:53.591337958 +0000 UTC m=+1032.203038181" watchObservedRunningTime="2025-12-03 16:20:53.598639103 +0000 UTC m=+1032.210339336" Dec 03 16:20:53 crc kubenswrapper[4998]: I1203 16:20:53.765140 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5h65w" podStartSLOduration=2.732265948 podStartE2EDuration="48.76511632s" podCreationTimestamp="2025-12-03 16:20:05 +0000 UTC" firstStartedPulling="2025-12-03 16:20:06.985126817 +0000 UTC m=+985.596827040" lastFinishedPulling="2025-12-03 16:20:53.017977189 +0000 UTC m=+1031.629677412" observedRunningTime="2025-12-03 16:20:53.74811819 +0000 UTC m=+1032.359818413" watchObservedRunningTime="2025-12-03 16:20:53.76511632 +0000 UTC m=+1032.376816543" Dec 03 16:20:54 crc kubenswrapper[4998]: I1203 16:20:54.486299 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gzg25" event={"ID":"7c372d0c-4cac-4e4c-bc8f-409b5c151d53","Type":"ContainerStarted","Data":"31682c8ccf22e26972b30137aa168a1b5cf132bc629b98b9f4fced19c1115605"} Dec 03 16:20:54 crc kubenswrapper[4998]: I1203 16:20:54.489112 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-28xct" event={"ID":"79a1881c-77c7-4c5c-b2fc-b2d7d9c31496","Type":"ContainerStarted","Data":"fe1c93c577c7244d0f553b7f36abb6f1d9350b9c5b2a840e72bb8fbb5fcfa7a0"} Dec 03 16:20:54 crc kubenswrapper[4998]: I1203 16:20:54.490647 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2h2m5" event={"ID":"f23eb2b1-107d-4167-a695-4479df0ae92a","Type":"ContainerStarted","Data":"3f9ea9085257a15e6dbed55f438d6799d89a76b32e2aaf2e8ed01a364065a693"} Dec 03 16:20:54 crc kubenswrapper[4998]: I1203 16:20:54.491084 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2h2m5" Dec 03 16:20:54 crc kubenswrapper[4998]: I1203 16:20:54.493545 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-m4gc7" event={"ID":"67e57367-12f9-487d-b691-ca243a28b056","Type":"ContainerStarted","Data":"88b26d296ba981b06c03e985b1f542b2a212e88fb57640c4e060ae36d5b7590b"} Dec 03 16:20:54 crc kubenswrapper[4998]: E1203 16:20:54.495898 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7\\\"\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-hr74g" podUID="ade59116-1aaf-44e0-b47b-15bdc7b402f0" Dec 03 16:20:54 crc kubenswrapper[4998]: E1203 16:20:54.500904 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" podUID="42a2439a-98ee-4172-813c-9ce7109fb1c0" Dec 03 16:20:54 crc kubenswrapper[4998]: I1203 16:20:54.510627 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-gzg25" podStartSLOduration=36.189260205 podStartE2EDuration="49.510606589s" podCreationTimestamp="2025-12-03 16:20:05 +0000 UTC" firstStartedPulling="2025-12-03 16:20:07.08062866 +0000 UTC m=+985.692328873" lastFinishedPulling="2025-12-03 16:20:20.401975034 +0000 UTC m=+999.013675257" observedRunningTime="2025-12-03 16:20:54.504203907 +0000 UTC m=+1033.115904140" watchObservedRunningTime="2025-12-03 16:20:54.510606589 +0000 UTC m=+1033.122306812" Dec 03 16:20:54 crc kubenswrapper[4998]: I1203 16:20:54.586008 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2h2m5" podStartSLOduration=3.509620903 podStartE2EDuration="49.585984254s" podCreationTimestamp="2025-12-03 16:20:05 +0000 UTC" firstStartedPulling="2025-12-03 16:20:07.223969333 +0000 UTC m=+985.835669556" lastFinishedPulling="2025-12-03 16:20:53.300332684 +0000 UTC m=+1031.912032907" observedRunningTime="2025-12-03 16:20:54.581231193 +0000 UTC m=+1033.192931436" watchObservedRunningTime="2025-12-03 16:20:54.585984254 +0000 UTC m=+1033.197684497" Dec 03 16:20:55 crc kubenswrapper[4998]: E1203 16:20:55.205567 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hzl7f" podUID="0e92e441-8150-4817-8039-9cdc8424e562" Dec 03 16:20:55 crc kubenswrapper[4998]: I1203 16:20:55.501976 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-tkz2j" event={"ID":"624bac01-c09f-424f-9000-feb138765cc4","Type":"ContainerStarted","Data":"5617a421bb7831693d74bd22829cb0b64b46dfd3dec387c36723915f42d59aea"} Dec 03 16:20:55 crc kubenswrapper[4998]: I1203 16:20:55.504914 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-ph8r8" event={"ID":"d926336a-658f-49bd-81b9-8086c8bb97a5","Type":"ContainerStarted","Data":"3cacd141a259d59a27b131ba185c61e2c139feb78cbc7c070dc683da33ea2760"} Dec 03 16:20:55 crc kubenswrapper[4998]: I1203 16:20:55.507591 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z8g4g" event={"ID":"6b89567b-4963-4113-9900-a132d873d260","Type":"ContainerStarted","Data":"feb9a7db3b010db0ca342ab45ebc6051da073f999ec85bae10202531703e683a"} Dec 03 16:20:55 crc kubenswrapper[4998]: I1203 16:20:55.507624 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z8g4g" event={"ID":"6b89567b-4963-4113-9900-a132d873d260","Type":"ContainerStarted","Data":"255ede09b4a894aaf936826f2ff3b4d55dc2daa61c24400d9fbaa1dbb5afd660"} Dec 03 16:20:55 crc kubenswrapper[4998]: I1203 16:20:55.509431 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-f6vdv" event={"ID":"aa3ac0c7-ea42-4728-8406-0ce3f9a96c21","Type":"ContainerStarted","Data":"4cc82eac686bc8f45de66f250564a9f58e02fd337222f6d708b95c0f602dbcc1"} Dec 03 16:20:55 crc kubenswrapper[4998]: I1203 16:20:55.511350 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-h72w5" event={"ID":"77fb606c-dea3-4e34-9f4c-5bb305e3b7f4","Type":"ContainerStarted","Data":"874ab7afee6cac8255ca49af140dc55fb2e665aa348570ac34d2358205efd03d"} Dec 03 16:20:55 crc kubenswrapper[4998]: I1203 16:20:55.513822 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-mn69l" event={"ID":"c92de5b9-8ef5-4f4c-b881-82c5f05cd853","Type":"ContainerStarted","Data":"5877190208b466b6e5e8374d1eca7b5802e26876dd3c935f25cfb60ae4330266"} Dec 03 16:20:55 crc kubenswrapper[4998]: I1203 16:20:55.516303 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hzl7f" event={"ID":"0e92e441-8150-4817-8039-9cdc8424e562","Type":"ContainerStarted","Data":"aa77af9d4dbf9fd7815161dc55cda0842867d4b142c0677fe93ce4e6d5a3f0cc"} Dec 03 16:20:55 crc kubenswrapper[4998]: I1203 16:20:55.518499 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-tv6bd" event={"ID":"a67cb778-507e-4dba-9ae4-60ae60315a0a","Type":"ContainerStarted","Data":"7382e56212ca66308bb1d3d2abb03e52c3b6bfb5e6e9aba39f14d23ce3267315"} Dec 03 16:20:55 crc kubenswrapper[4998]: E1203 16:20:55.520208 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hzl7f" podUID="0e92e441-8150-4817-8039-9cdc8424e562" Dec 03 16:20:55 crc kubenswrapper[4998]: E1203 16:20:55.551590 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2wrqc" podUID="d7adeced-3036-49a7-8806-fdbefce6406f" Dec 03 16:20:55 crc kubenswrapper[4998]: I1203 16:20:55.569193 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-tkz2j" podStartSLOduration=36.365722484 podStartE2EDuration="50.569167389s" podCreationTimestamp="2025-12-03 16:20:05 +0000 UTC" firstStartedPulling="2025-12-03 16:20:06.184408492 +0000 UTC m=+984.796108715" lastFinishedPulling="2025-12-03 16:20:20.387853357 +0000 UTC m=+998.999553620" observedRunningTime="2025-12-03 16:20:55.534149324 +0000 UTC m=+1034.145849547" watchObservedRunningTime="2025-12-03 16:20:55.569167389 +0000 UTC m=+1034.180867612" Dec 03 16:20:55 crc kubenswrapper[4998]: I1203 16:20:55.600637 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-28xct" podStartSLOduration=4.43198075 podStartE2EDuration="50.600613634s" podCreationTimestamp="2025-12-03 16:20:05 +0000 UTC" firstStartedPulling="2025-12-03 16:20:07.279390953 +0000 UTC m=+985.891091176" lastFinishedPulling="2025-12-03 16:20:53.448023837 +0000 UTC m=+1032.059724060" observedRunningTime="2025-12-03 16:20:55.571780155 +0000 UTC m=+1034.183480398" watchObservedRunningTime="2025-12-03 16:20:55.600613634 +0000 UTC m=+1034.212313857" Dec 03 16:20:55 crc kubenswrapper[4998]: I1203 16:20:55.602960 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-ph8r8" podStartSLOduration=37.28596823 podStartE2EDuration="50.602946783s" podCreationTimestamp="2025-12-03 16:20:05 +0000 UTC" firstStartedPulling="2025-12-03 16:20:07.07982196 +0000 UTC m=+985.691522173" lastFinishedPulling="2025-12-03 16:20:20.396800493 +0000 UTC m=+999.008500726" observedRunningTime="2025-12-03 16:20:55.596651904 +0000 UTC m=+1034.208352127" watchObservedRunningTime="2025-12-03 16:20:55.602946783 +0000 UTC m=+1034.214647006" Dec 03 16:20:55 crc kubenswrapper[4998]: I1203 16:20:55.624434 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-f6vdv" podStartSLOduration=37.264772274 podStartE2EDuration="50.624419115s" podCreationTimestamp="2025-12-03 16:20:05 +0000 UTC" firstStartedPulling="2025-12-03 16:20:07.071973932 +0000 UTC m=+985.683674155" lastFinishedPulling="2025-12-03 16:20:20.431620763 +0000 UTC m=+999.043320996" observedRunningTime="2025-12-03 16:20:55.619917192 +0000 UTC m=+1034.231617445" watchObservedRunningTime="2025-12-03 16:20:55.624419115 +0000 UTC m=+1034.236119338" Dec 03 16:20:55 crc kubenswrapper[4998]: I1203 16:20:55.656111 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-mn69l" podStartSLOduration=36.832666625 podStartE2EDuration="50.656089976s" podCreationTimestamp="2025-12-03 16:20:05 +0000 UTC" firstStartedPulling="2025-12-03 16:20:06.589040808 +0000 UTC m=+985.200741031" lastFinishedPulling="2025-12-03 16:20:20.412464159 +0000 UTC m=+999.024164382" observedRunningTime="2025-12-03 16:20:55.651858309 +0000 UTC m=+1034.263558542" watchObservedRunningTime="2025-12-03 16:20:55.656089976 +0000 UTC m=+1034.267790199" Dec 03 16:20:55 crc kubenswrapper[4998]: I1203 16:20:55.658606 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-tcltm" Dec 03 16:20:55 crc kubenswrapper[4998]: I1203 16:20:55.671157 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-tcltm" Dec 03 16:20:55 crc kubenswrapper[4998]: I1203 16:20:55.716679 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-h72w5" podStartSLOduration=37.381152024 podStartE2EDuration="50.716656536s" podCreationTimestamp="2025-12-03 16:20:05 +0000 UTC" firstStartedPulling="2025-12-03 16:20:07.074885485 +0000 UTC m=+985.686585708" lastFinishedPulling="2025-12-03 16:20:20.410389957 +0000 UTC m=+999.022090220" observedRunningTime="2025-12-03 16:20:55.710989813 +0000 UTC m=+1034.322690026" watchObservedRunningTime="2025-12-03 16:20:55.716656536 +0000 UTC m=+1034.328356769" Dec 03 16:20:55 crc kubenswrapper[4998]: I1203 16:20:55.744294 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-tv6bd" podStartSLOduration=37.261385708 podStartE2EDuration="50.744278944s" podCreationTimestamp="2025-12-03 16:20:05 +0000 UTC" firstStartedPulling="2025-12-03 16:20:06.928635179 +0000 UTC m=+985.540335402" lastFinishedPulling="2025-12-03 16:20:20.411528415 +0000 UTC m=+999.023228638" observedRunningTime="2025-12-03 16:20:55.73422215 +0000 UTC m=+1034.345922373" watchObservedRunningTime="2025-12-03 16:20:55.744278944 +0000 UTC m=+1034.355979167" Dec 03 16:20:55 crc kubenswrapper[4998]: I1203 16:20:55.780892 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-m4gc7" podStartSLOduration=37.282482841 podStartE2EDuration="50.780874479s" podCreationTimestamp="2025-12-03 16:20:05 +0000 UTC" firstStartedPulling="2025-12-03 16:20:06.889455409 +0000 UTC m=+985.501155632" lastFinishedPulling="2025-12-03 16:20:20.387847047 +0000 UTC m=+998.999547270" observedRunningTime="2025-12-03 16:20:55.779197577 +0000 UTC m=+1034.390897800" watchObservedRunningTime="2025-12-03 16:20:55.780874479 +0000 UTC m=+1034.392574702" Dec 03 16:20:55 crc kubenswrapper[4998]: I1203 16:20:55.856080 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d2gcr" Dec 03 16:20:55 crc kubenswrapper[4998]: I1203 16:20:55.867676 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d2gcr" Dec 03 16:20:56 crc kubenswrapper[4998]: I1203 16:20:56.095391 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-28xct" Dec 03 16:20:56 crc kubenswrapper[4998]: I1203 16:20:56.526850 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d2gcr" event={"ID":"e52d7801-a5c9-47a3-bedc-27174cfbd993","Type":"ContainerStarted","Data":"be78e89f98e753adf7351dc48d990bcbb91e95c6e9d968a75a0f0e8565e2ab89"} Dec 03 16:20:56 crc kubenswrapper[4998]: I1203 16:20:56.529643 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2wrqc" event={"ID":"d7adeced-3036-49a7-8806-fdbefce6406f","Type":"ContainerStarted","Data":"4dda29ce1cfbf10217ca1f39b56b2b0170137c2b86760df8a618dd01219d59a9"} Dec 03 16:20:56 crc kubenswrapper[4998]: E1203 16:20:56.531519 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2wrqc" podUID="d7adeced-3036-49a7-8806-fdbefce6406f" Dec 03 16:20:56 crc kubenswrapper[4998]: I1203 16:20:56.535121 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-tcltm" event={"ID":"17cdd58c-7a0c-42e0-bcbe-4545050637e3","Type":"ContainerStarted","Data":"1e119a3f46af60c54b289922183895f6460e9d1bbbafb82cc70d4107146301d9"} Dec 03 16:20:56 crc kubenswrapper[4998]: I1203 16:20:56.547634 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-d2gcr" podStartSLOduration=38.207710552 podStartE2EDuration="51.547616034s" podCreationTimestamp="2025-12-03 16:20:05 +0000 UTC" firstStartedPulling="2025-12-03 16:20:07.074630239 +0000 UTC m=+985.686330462" lastFinishedPulling="2025-12-03 16:20:20.414535721 +0000 UTC m=+999.026235944" observedRunningTime="2025-12-03 16:20:56.546130177 +0000 UTC m=+1035.157830410" watchObservedRunningTime="2025-12-03 16:20:56.547616034 +0000 UTC m=+1035.159316257" Dec 03 16:20:56 crc kubenswrapper[4998]: I1203 16:20:56.613200 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z8g4g" podStartSLOduration=4.074909306 podStartE2EDuration="51.613182161s" podCreationTimestamp="2025-12-03 16:20:05 +0000 UTC" firstStartedPulling="2025-12-03 16:20:07.211841206 +0000 UTC m=+985.823541439" lastFinishedPulling="2025-12-03 16:20:54.750114071 +0000 UTC m=+1033.361814294" observedRunningTime="2025-12-03 16:20:56.6012703 +0000 UTC m=+1035.212970513" watchObservedRunningTime="2025-12-03 16:20:56.613182161 +0000 UTC m=+1035.224882384" Dec 03 16:20:56 crc kubenswrapper[4998]: I1203 16:20:56.634570 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-tcltm" podStartSLOduration=37.533664748 podStartE2EDuration="51.634549531s" podCreationTimestamp="2025-12-03 16:20:05 +0000 UTC" firstStartedPulling="2025-12-03 16:20:06.33035204 +0000 UTC m=+984.942052263" lastFinishedPulling="2025-12-03 16:20:20.431236823 +0000 UTC m=+999.042937046" observedRunningTime="2025-12-03 16:20:56.628354434 +0000 UTC m=+1035.240054657" watchObservedRunningTime="2025-12-03 16:20:56.634549531 +0000 UTC m=+1035.246249754" Dec 03 16:20:58 crc kubenswrapper[4998]: I1203 16:20:58.184136 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-67847ddbd8-t8wbd" Dec 03 16:21:04 crc kubenswrapper[4998]: E1203 16:21:04.680293 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.32:5001/openstack-k8s-operators/watcher-operator:164610071c31587da7ddb7356abc05e8c519dbfa\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-f77f967dc-ndj9c" podUID="ac213ee6-009b-4098-9d18-6dfc8cb97a0e" Dec 03 16:21:04 crc kubenswrapper[4998]: E1203 16:21:04.680287 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fqdqg" podUID="9f587468-9ae1-43af-8652-f4394865ff37" Dec 03 16:21:05 crc kubenswrapper[4998]: I1203 16:21:05.988600 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2h2m5" Dec 03 16:21:06 crc kubenswrapper[4998]: I1203 16:21:06.099706 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-28xct" Dec 03 16:21:06 crc kubenswrapper[4998]: I1203 16:21:06.320471 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z8g4g" Dec 03 16:21:06 crc kubenswrapper[4998]: I1203 16:21:06.322598 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-z8g4g" Dec 03 16:21:06 crc kubenswrapper[4998]: E1203 16:21:06.679277 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hzl7f" podUID="0e92e441-8150-4817-8039-9cdc8424e562" Dec 03 16:21:11 crc kubenswrapper[4998]: I1203 16:21:11.661620 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-hr74g" event={"ID":"ade59116-1aaf-44e0-b47b-15bdc7b402f0","Type":"ContainerStarted","Data":"bef2df85d8500a42dd74eb2195e87bdf5380065217da9881a2f897fcffaa0a81"} Dec 03 16:21:11 crc kubenswrapper[4998]: I1203 16:21:11.662454 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-hr74g" Dec 03 16:21:11 crc kubenswrapper[4998]: I1203 16:21:11.664436 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" event={"ID":"42a2439a-98ee-4172-813c-9ce7109fb1c0","Type":"ContainerStarted","Data":"adf6d6a23094c9d2c6ce5d72bee206ba1e7248f76e8220d4f6903fa347c3bc3d"} Dec 03 16:21:11 crc kubenswrapper[4998]: I1203 16:21:11.664669 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" Dec 03 16:21:11 crc kubenswrapper[4998]: I1203 16:21:11.690919 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-hr74g" podStartSLOduration=18.74566249 podStartE2EDuration="1m6.690898519s" podCreationTimestamp="2025-12-03 16:20:05 +0000 UTC" firstStartedPulling="2025-12-03 16:20:22.5074427 +0000 UTC m=+1001.119142913" lastFinishedPulling="2025-12-03 16:21:10.452678709 +0000 UTC m=+1049.064378942" observedRunningTime="2025-12-03 16:21:11.688588051 +0000 UTC m=+1050.300288284" watchObservedRunningTime="2025-12-03 16:21:11.690898519 +0000 UTC m=+1050.302598742" Dec 03 16:21:11 crc kubenswrapper[4998]: I1203 16:21:11.724272 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" podStartSLOduration=18.248863746 podStartE2EDuration="1m6.724255382s" podCreationTimestamp="2025-12-03 16:20:05 +0000 UTC" firstStartedPulling="2025-12-03 16:20:22.548858287 +0000 UTC m=+1001.160558510" lastFinishedPulling="2025-12-03 16:21:11.024249923 +0000 UTC m=+1049.635950146" observedRunningTime="2025-12-03 16:21:11.718733613 +0000 UTC m=+1050.330433836" watchObservedRunningTime="2025-12-03 16:21:11.724255382 +0000 UTC m=+1050.335955605" Dec 03 16:21:12 crc kubenswrapper[4998]: I1203 16:21:12.676715 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2wrqc" event={"ID":"d7adeced-3036-49a7-8806-fdbefce6406f","Type":"ContainerStarted","Data":"a0971dd0b30bb37945187a21537b8a95fe6ec0c3d8beb5a5d7e10de6a4425c9c"} Dec 03 16:21:12 crc kubenswrapper[4998]: I1203 16:21:12.677313 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2wrqc" Dec 03 16:21:12 crc kubenswrapper[4998]: I1203 16:21:12.703445 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2wrqc" podStartSLOduration=2.820462066 podStartE2EDuration="1m7.703419246s" podCreationTimestamp="2025-12-03 16:20:05 +0000 UTC" firstStartedPulling="2025-12-03 16:20:07.275432363 +0000 UTC m=+985.887132586" lastFinishedPulling="2025-12-03 16:21:12.158389523 +0000 UTC m=+1050.770089766" observedRunningTime="2025-12-03 16:21:12.698056461 +0000 UTC m=+1051.309756704" watchObservedRunningTime="2025-12-03 16:21:12.703419246 +0000 UTC m=+1051.315119489" Dec 03 16:21:16 crc kubenswrapper[4998]: I1203 16:21:16.711101 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-f77f967dc-ndj9c" event={"ID":"ac213ee6-009b-4098-9d18-6dfc8cb97a0e","Type":"ContainerStarted","Data":"2d2df326e9ace85130f13d3d0d4fd0d125e136ad4618a73c3f59e299cf33e5ae"} Dec 03 16:21:16 crc kubenswrapper[4998]: I1203 16:21:16.711938 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-f77f967dc-ndj9c" Dec 03 16:21:16 crc kubenswrapper[4998]: I1203 16:21:16.736442 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-f77f967dc-ndj9c" podStartSLOduration=2.8521940580000003 podStartE2EDuration="1m11.736411541s" podCreationTimestamp="2025-12-03 16:20:05 +0000 UTC" firstStartedPulling="2025-12-03 16:20:07.231277987 +0000 UTC m=+985.842978210" lastFinishedPulling="2025-12-03 16:21:16.11549546 +0000 UTC m=+1054.727195693" observedRunningTime="2025-12-03 16:21:16.731303952 +0000 UTC m=+1055.343004175" watchObservedRunningTime="2025-12-03 16:21:16.736411541 +0000 UTC m=+1055.348111814" Dec 03 16:21:17 crc kubenswrapper[4998]: I1203 16:21:17.723244 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fqdqg" event={"ID":"9f587468-9ae1-43af-8652-f4394865ff37","Type":"ContainerStarted","Data":"50e46a70852353306b89ade912ab5410bebaec94514da37bc28c13004ce74954"} Dec 03 16:21:17 crc kubenswrapper[4998]: I1203 16:21:17.742845 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-fqdqg" podStartSLOduration=3.146821913 podStartE2EDuration="1m12.742820813s" podCreationTimestamp="2025-12-03 16:20:05 +0000 UTC" firstStartedPulling="2025-12-03 16:20:07.262046905 +0000 UTC m=+985.873747128" lastFinishedPulling="2025-12-03 16:21:16.858045805 +0000 UTC m=+1055.469746028" observedRunningTime="2025-12-03 16:21:17.741812127 +0000 UTC m=+1056.353512420" watchObservedRunningTime="2025-12-03 16:21:17.742820813 +0000 UTC m=+1056.354521076" Dec 03 16:21:18 crc kubenswrapper[4998]: I1203 16:21:18.730788 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hzl7f" event={"ID":"0e92e441-8150-4817-8039-9cdc8424e562","Type":"ContainerStarted","Data":"d6ff3a99bd5d480f20b333cb588039b7d7412f76b752d62bf9474aa5524a5a6c"} Dec 03 16:21:18 crc kubenswrapper[4998]: I1203 16:21:18.731225 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hzl7f" Dec 03 16:21:18 crc kubenswrapper[4998]: I1203 16:21:18.749638 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hzl7f" podStartSLOduration=2.92747724 podStartE2EDuration="1m13.749616555s" podCreationTimestamp="2025-12-03 16:20:05 +0000 UTC" firstStartedPulling="2025-12-03 16:20:07.275166246 +0000 UTC m=+985.886866469" lastFinishedPulling="2025-12-03 16:21:18.097305561 +0000 UTC m=+1056.709005784" observedRunningTime="2025-12-03 16:21:18.745601893 +0000 UTC m=+1057.357302126" watchObservedRunningTime="2025-12-03 16:21:18.749616555 +0000 UTC m=+1057.361316778" Dec 03 16:21:21 crc kubenswrapper[4998]: I1203 16:21:21.216637 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-hr74g" Dec 03 16:21:21 crc kubenswrapper[4998]: I1203 16:21:21.657375 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz" Dec 03 16:21:25 crc kubenswrapper[4998]: I1203 16:21:25.964937 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-hzl7f" Dec 03 16:21:26 crc kubenswrapper[4998]: I1203 16:21:26.190844 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-2wrqc" Dec 03 16:21:26 crc kubenswrapper[4998]: I1203 16:21:26.342645 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-f77f967dc-ndj9c" Dec 03 16:21:27 crc kubenswrapper[4998]: I1203 16:21:27.110790 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:21:27 crc kubenswrapper[4998]: I1203 16:21:27.110848 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:21:44 crc kubenswrapper[4998]: I1203 16:21:44.920399 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-548dc7894c-6vwpc"] Dec 03 16:21:44 crc kubenswrapper[4998]: I1203 16:21:44.922091 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-548dc7894c-6vwpc" Dec 03 16:21:44 crc kubenswrapper[4998]: I1203 16:21:44.926887 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 03 16:21:44 crc kubenswrapper[4998]: I1203 16:21:44.928228 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 03 16:21:44 crc kubenswrapper[4998]: I1203 16:21:44.928460 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 03 16:21:44 crc kubenswrapper[4998]: I1203 16:21:44.930451 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-2wccc" Dec 03 16:21:44 crc kubenswrapper[4998]: I1203 16:21:44.937429 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-548dc7894c-6vwpc"] Dec 03 16:21:45 crc kubenswrapper[4998]: I1203 16:21:45.000297 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77b567bfc7-g5znf"] Dec 03 16:21:45 crc kubenswrapper[4998]: I1203 16:21:45.001499 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77b567bfc7-g5znf" Dec 03 16:21:45 crc kubenswrapper[4998]: I1203 16:21:45.006993 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 03 16:21:45 crc kubenswrapper[4998]: I1203 16:21:45.008481 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77b567bfc7-g5znf"] Dec 03 16:21:45 crc kubenswrapper[4998]: I1203 16:21:45.038214 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmfd9\" (UniqueName: \"kubernetes.io/projected/3e93a28e-5430-4fac-8948-67ddf8db3a87-kube-api-access-fmfd9\") pod \"dnsmasq-dns-548dc7894c-6vwpc\" (UID: \"3e93a28e-5430-4fac-8948-67ddf8db3a87\") " pod="openstack/dnsmasq-dns-548dc7894c-6vwpc" Dec 03 16:21:45 crc kubenswrapper[4998]: I1203 16:21:45.038298 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e93a28e-5430-4fac-8948-67ddf8db3a87-config\") pod \"dnsmasq-dns-548dc7894c-6vwpc\" (UID: \"3e93a28e-5430-4fac-8948-67ddf8db3a87\") " pod="openstack/dnsmasq-dns-548dc7894c-6vwpc" Dec 03 16:21:45 crc kubenswrapper[4998]: I1203 16:21:45.140088 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmfd9\" (UniqueName: \"kubernetes.io/projected/3e93a28e-5430-4fac-8948-67ddf8db3a87-kube-api-access-fmfd9\") pod \"dnsmasq-dns-548dc7894c-6vwpc\" (UID: \"3e93a28e-5430-4fac-8948-67ddf8db3a87\") " pod="openstack/dnsmasq-dns-548dc7894c-6vwpc" Dec 03 16:21:45 crc kubenswrapper[4998]: I1203 16:21:45.140168 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwkfk\" (UniqueName: \"kubernetes.io/projected/8d2d7661-f999-4c89-9e97-dce07062c850-kube-api-access-kwkfk\") pod \"dnsmasq-dns-77b567bfc7-g5znf\" (UID: \"8d2d7661-f999-4c89-9e97-dce07062c850\") " pod="openstack/dnsmasq-dns-77b567bfc7-g5znf" Dec 03 16:21:45 crc kubenswrapper[4998]: I1203 16:21:45.140196 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e93a28e-5430-4fac-8948-67ddf8db3a87-config\") pod \"dnsmasq-dns-548dc7894c-6vwpc\" (UID: \"3e93a28e-5430-4fac-8948-67ddf8db3a87\") " pod="openstack/dnsmasq-dns-548dc7894c-6vwpc" Dec 03 16:21:45 crc kubenswrapper[4998]: I1203 16:21:45.140221 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d2d7661-f999-4c89-9e97-dce07062c850-config\") pod \"dnsmasq-dns-77b567bfc7-g5znf\" (UID: \"8d2d7661-f999-4c89-9e97-dce07062c850\") " pod="openstack/dnsmasq-dns-77b567bfc7-g5znf" Dec 03 16:21:45 crc kubenswrapper[4998]: I1203 16:21:45.140244 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d2d7661-f999-4c89-9e97-dce07062c850-dns-svc\") pod \"dnsmasq-dns-77b567bfc7-g5znf\" (UID: \"8d2d7661-f999-4c89-9e97-dce07062c850\") " pod="openstack/dnsmasq-dns-77b567bfc7-g5znf" Dec 03 16:21:45 crc kubenswrapper[4998]: I1203 16:21:45.141208 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e93a28e-5430-4fac-8948-67ddf8db3a87-config\") pod \"dnsmasq-dns-548dc7894c-6vwpc\" (UID: \"3e93a28e-5430-4fac-8948-67ddf8db3a87\") " pod="openstack/dnsmasq-dns-548dc7894c-6vwpc" Dec 03 16:21:45 crc kubenswrapper[4998]: I1203 16:21:45.163733 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmfd9\" (UniqueName: \"kubernetes.io/projected/3e93a28e-5430-4fac-8948-67ddf8db3a87-kube-api-access-fmfd9\") pod \"dnsmasq-dns-548dc7894c-6vwpc\" (UID: \"3e93a28e-5430-4fac-8948-67ddf8db3a87\") " pod="openstack/dnsmasq-dns-548dc7894c-6vwpc" Dec 03 16:21:45 crc kubenswrapper[4998]: I1203 16:21:45.241213 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-548dc7894c-6vwpc" Dec 03 16:21:45 crc kubenswrapper[4998]: I1203 16:21:45.241567 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwkfk\" (UniqueName: \"kubernetes.io/projected/8d2d7661-f999-4c89-9e97-dce07062c850-kube-api-access-kwkfk\") pod \"dnsmasq-dns-77b567bfc7-g5znf\" (UID: \"8d2d7661-f999-4c89-9e97-dce07062c850\") " pod="openstack/dnsmasq-dns-77b567bfc7-g5znf" Dec 03 16:21:45 crc kubenswrapper[4998]: I1203 16:21:45.241624 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d2d7661-f999-4c89-9e97-dce07062c850-config\") pod \"dnsmasq-dns-77b567bfc7-g5znf\" (UID: \"8d2d7661-f999-4c89-9e97-dce07062c850\") " pod="openstack/dnsmasq-dns-77b567bfc7-g5znf" Dec 03 16:21:45 crc kubenswrapper[4998]: I1203 16:21:45.241650 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d2d7661-f999-4c89-9e97-dce07062c850-dns-svc\") pod \"dnsmasq-dns-77b567bfc7-g5znf\" (UID: \"8d2d7661-f999-4c89-9e97-dce07062c850\") " pod="openstack/dnsmasq-dns-77b567bfc7-g5znf" Dec 03 16:21:45 crc kubenswrapper[4998]: I1203 16:21:45.242439 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d2d7661-f999-4c89-9e97-dce07062c850-dns-svc\") pod \"dnsmasq-dns-77b567bfc7-g5znf\" (UID: \"8d2d7661-f999-4c89-9e97-dce07062c850\") " pod="openstack/dnsmasq-dns-77b567bfc7-g5znf" Dec 03 16:21:45 crc kubenswrapper[4998]: I1203 16:21:45.242495 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d2d7661-f999-4c89-9e97-dce07062c850-config\") pod \"dnsmasq-dns-77b567bfc7-g5znf\" (UID: \"8d2d7661-f999-4c89-9e97-dce07062c850\") " pod="openstack/dnsmasq-dns-77b567bfc7-g5znf" Dec 03 16:21:45 crc kubenswrapper[4998]: I1203 16:21:45.279265 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwkfk\" (UniqueName: \"kubernetes.io/projected/8d2d7661-f999-4c89-9e97-dce07062c850-kube-api-access-kwkfk\") pod \"dnsmasq-dns-77b567bfc7-g5znf\" (UID: \"8d2d7661-f999-4c89-9e97-dce07062c850\") " pod="openstack/dnsmasq-dns-77b567bfc7-g5znf" Dec 03 16:21:45 crc kubenswrapper[4998]: I1203 16:21:45.319831 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77b567bfc7-g5znf" Dec 03 16:21:45 crc kubenswrapper[4998]: I1203 16:21:45.611953 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77b567bfc7-g5znf"] Dec 03 16:21:45 crc kubenswrapper[4998]: W1203 16:21:45.615724 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d2d7661_f999_4c89_9e97_dce07062c850.slice/crio-88caefdb3437a280bdcd0dd9e91b9522cb285d60db2623e2f92e459a2ad3fc44 WatchSource:0}: Error finding container 88caefdb3437a280bdcd0dd9e91b9522cb285d60db2623e2f92e459a2ad3fc44: Status 404 returned error can't find the container with id 88caefdb3437a280bdcd0dd9e91b9522cb285d60db2623e2f92e459a2ad3fc44 Dec 03 16:21:45 crc kubenswrapper[4998]: I1203 16:21:45.730405 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-548dc7894c-6vwpc"] Dec 03 16:21:45 crc kubenswrapper[4998]: W1203 16:21:45.741861 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e93a28e_5430_4fac_8948_67ddf8db3a87.slice/crio-4b2d63cfa91e30b46976fa9b56c9da82777e3d470ec0f093788aeabd6e7a0783 WatchSource:0}: Error finding container 4b2d63cfa91e30b46976fa9b56c9da82777e3d470ec0f093788aeabd6e7a0783: Status 404 returned error can't find the container with id 4b2d63cfa91e30b46976fa9b56c9da82777e3d470ec0f093788aeabd6e7a0783 Dec 03 16:21:45 crc kubenswrapper[4998]: I1203 16:21:45.979478 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-548dc7894c-6vwpc" event={"ID":"3e93a28e-5430-4fac-8948-67ddf8db3a87","Type":"ContainerStarted","Data":"4b2d63cfa91e30b46976fa9b56c9da82777e3d470ec0f093788aeabd6e7a0783"} Dec 03 16:21:45 crc kubenswrapper[4998]: I1203 16:21:45.980485 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77b567bfc7-g5znf" event={"ID":"8d2d7661-f999-4c89-9e97-dce07062c850","Type":"ContainerStarted","Data":"88caefdb3437a280bdcd0dd9e91b9522cb285d60db2623e2f92e459a2ad3fc44"} Dec 03 16:21:48 crc kubenswrapper[4998]: I1203 16:21:48.637280 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-548dc7894c-6vwpc"] Dec 03 16:21:48 crc kubenswrapper[4998]: I1203 16:21:48.663273 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76c44b4bf7-v5mhz"] Dec 03 16:21:48 crc kubenswrapper[4998]: I1203 16:21:48.664824 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76c44b4bf7-v5mhz" Dec 03 16:21:48 crc kubenswrapper[4998]: I1203 16:21:48.682801 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76c44b4bf7-v5mhz"] Dec 03 16:21:48 crc kubenswrapper[4998]: I1203 16:21:48.800414 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0c40a4e-c357-48d4-a3e1-a1eb8ba85256-config\") pod \"dnsmasq-dns-76c44b4bf7-v5mhz\" (UID: \"c0c40a4e-c357-48d4-a3e1-a1eb8ba85256\") " pod="openstack/dnsmasq-dns-76c44b4bf7-v5mhz" Dec 03 16:21:48 crc kubenswrapper[4998]: I1203 16:21:48.800488 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzvgk\" (UniqueName: \"kubernetes.io/projected/c0c40a4e-c357-48d4-a3e1-a1eb8ba85256-kube-api-access-kzvgk\") pod \"dnsmasq-dns-76c44b4bf7-v5mhz\" (UID: \"c0c40a4e-c357-48d4-a3e1-a1eb8ba85256\") " pod="openstack/dnsmasq-dns-76c44b4bf7-v5mhz" Dec 03 16:21:48 crc kubenswrapper[4998]: I1203 16:21:48.800627 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c0c40a4e-c357-48d4-a3e1-a1eb8ba85256-dns-svc\") pod \"dnsmasq-dns-76c44b4bf7-v5mhz\" (UID: \"c0c40a4e-c357-48d4-a3e1-a1eb8ba85256\") " pod="openstack/dnsmasq-dns-76c44b4bf7-v5mhz" Dec 03 16:21:48 crc kubenswrapper[4998]: I1203 16:21:48.902743 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c0c40a4e-c357-48d4-a3e1-a1eb8ba85256-dns-svc\") pod \"dnsmasq-dns-76c44b4bf7-v5mhz\" (UID: \"c0c40a4e-c357-48d4-a3e1-a1eb8ba85256\") " pod="openstack/dnsmasq-dns-76c44b4bf7-v5mhz" Dec 03 16:21:48 crc kubenswrapper[4998]: I1203 16:21:48.902830 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0c40a4e-c357-48d4-a3e1-a1eb8ba85256-config\") pod \"dnsmasq-dns-76c44b4bf7-v5mhz\" (UID: \"c0c40a4e-c357-48d4-a3e1-a1eb8ba85256\") " pod="openstack/dnsmasq-dns-76c44b4bf7-v5mhz" Dec 03 16:21:48 crc kubenswrapper[4998]: I1203 16:21:48.902868 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzvgk\" (UniqueName: \"kubernetes.io/projected/c0c40a4e-c357-48d4-a3e1-a1eb8ba85256-kube-api-access-kzvgk\") pod \"dnsmasq-dns-76c44b4bf7-v5mhz\" (UID: \"c0c40a4e-c357-48d4-a3e1-a1eb8ba85256\") " pod="openstack/dnsmasq-dns-76c44b4bf7-v5mhz" Dec 03 16:21:48 crc kubenswrapper[4998]: I1203 16:21:48.904122 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c0c40a4e-c357-48d4-a3e1-a1eb8ba85256-dns-svc\") pod \"dnsmasq-dns-76c44b4bf7-v5mhz\" (UID: \"c0c40a4e-c357-48d4-a3e1-a1eb8ba85256\") " pod="openstack/dnsmasq-dns-76c44b4bf7-v5mhz" Dec 03 16:21:48 crc kubenswrapper[4998]: I1203 16:21:48.904545 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0c40a4e-c357-48d4-a3e1-a1eb8ba85256-config\") pod \"dnsmasq-dns-76c44b4bf7-v5mhz\" (UID: \"c0c40a4e-c357-48d4-a3e1-a1eb8ba85256\") " pod="openstack/dnsmasq-dns-76c44b4bf7-v5mhz" Dec 03 16:21:48 crc kubenswrapper[4998]: I1203 16:21:48.928313 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzvgk\" (UniqueName: \"kubernetes.io/projected/c0c40a4e-c357-48d4-a3e1-a1eb8ba85256-kube-api-access-kzvgk\") pod \"dnsmasq-dns-76c44b4bf7-v5mhz\" (UID: \"c0c40a4e-c357-48d4-a3e1-a1eb8ba85256\") " pod="openstack/dnsmasq-dns-76c44b4bf7-v5mhz" Dec 03 16:21:48 crc kubenswrapper[4998]: I1203 16:21:48.956534 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77b567bfc7-g5znf"] Dec 03 16:21:48 crc kubenswrapper[4998]: I1203 16:21:48.971680 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7786d8fd7-fmq4c"] Dec 03 16:21:48 crc kubenswrapper[4998]: I1203 16:21:48.972851 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7786d8fd7-fmq4c" Dec 03 16:21:48 crc kubenswrapper[4998]: I1203 16:21:48.981173 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7786d8fd7-fmq4c"] Dec 03 16:21:48 crc kubenswrapper[4998]: I1203 16:21:48.985516 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76c44b4bf7-v5mhz" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.105334 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5pdb\" (UniqueName: \"kubernetes.io/projected/27c5a979-99f8-46a9-9a96-312c8c61186f-kube-api-access-v5pdb\") pod \"dnsmasq-dns-7786d8fd7-fmq4c\" (UID: \"27c5a979-99f8-46a9-9a96-312c8c61186f\") " pod="openstack/dnsmasq-dns-7786d8fd7-fmq4c" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.105394 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27c5a979-99f8-46a9-9a96-312c8c61186f-config\") pod \"dnsmasq-dns-7786d8fd7-fmq4c\" (UID: \"27c5a979-99f8-46a9-9a96-312c8c61186f\") " pod="openstack/dnsmasq-dns-7786d8fd7-fmq4c" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.105431 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27c5a979-99f8-46a9-9a96-312c8c61186f-dns-svc\") pod \"dnsmasq-dns-7786d8fd7-fmq4c\" (UID: \"27c5a979-99f8-46a9-9a96-312c8c61186f\") " pod="openstack/dnsmasq-dns-7786d8fd7-fmq4c" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.206495 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27c5a979-99f8-46a9-9a96-312c8c61186f-dns-svc\") pod \"dnsmasq-dns-7786d8fd7-fmq4c\" (UID: \"27c5a979-99f8-46a9-9a96-312c8c61186f\") " pod="openstack/dnsmasq-dns-7786d8fd7-fmq4c" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.206604 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5pdb\" (UniqueName: \"kubernetes.io/projected/27c5a979-99f8-46a9-9a96-312c8c61186f-kube-api-access-v5pdb\") pod \"dnsmasq-dns-7786d8fd7-fmq4c\" (UID: \"27c5a979-99f8-46a9-9a96-312c8c61186f\") " pod="openstack/dnsmasq-dns-7786d8fd7-fmq4c" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.206647 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27c5a979-99f8-46a9-9a96-312c8c61186f-config\") pod \"dnsmasq-dns-7786d8fd7-fmq4c\" (UID: \"27c5a979-99f8-46a9-9a96-312c8c61186f\") " pod="openstack/dnsmasq-dns-7786d8fd7-fmq4c" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.207467 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27c5a979-99f8-46a9-9a96-312c8c61186f-dns-svc\") pod \"dnsmasq-dns-7786d8fd7-fmq4c\" (UID: \"27c5a979-99f8-46a9-9a96-312c8c61186f\") " pod="openstack/dnsmasq-dns-7786d8fd7-fmq4c" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.207509 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27c5a979-99f8-46a9-9a96-312c8c61186f-config\") pod \"dnsmasq-dns-7786d8fd7-fmq4c\" (UID: \"27c5a979-99f8-46a9-9a96-312c8c61186f\") " pod="openstack/dnsmasq-dns-7786d8fd7-fmq4c" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.234582 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5pdb\" (UniqueName: \"kubernetes.io/projected/27c5a979-99f8-46a9-9a96-312c8c61186f-kube-api-access-v5pdb\") pod \"dnsmasq-dns-7786d8fd7-fmq4c\" (UID: \"27c5a979-99f8-46a9-9a96-312c8c61186f\") " pod="openstack/dnsmasq-dns-7786d8fd7-fmq4c" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.293919 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7786d8fd7-fmq4c" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.328389 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76c44b4bf7-v5mhz"] Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.347136 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-594b65fc49-cgfb2"] Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.348250 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-594b65fc49-cgfb2" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.370575 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-594b65fc49-cgfb2"] Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.412597 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39e12872-2efe-4cec-82dd-a265f58da9ec-dns-svc\") pod \"dnsmasq-dns-594b65fc49-cgfb2\" (UID: \"39e12872-2efe-4cec-82dd-a265f58da9ec\") " pod="openstack/dnsmasq-dns-594b65fc49-cgfb2" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.412687 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39e12872-2efe-4cec-82dd-a265f58da9ec-config\") pod \"dnsmasq-dns-594b65fc49-cgfb2\" (UID: \"39e12872-2efe-4cec-82dd-a265f58da9ec\") " pod="openstack/dnsmasq-dns-594b65fc49-cgfb2" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.412809 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9j62\" (UniqueName: \"kubernetes.io/projected/39e12872-2efe-4cec-82dd-a265f58da9ec-kube-api-access-s9j62\") pod \"dnsmasq-dns-594b65fc49-cgfb2\" (UID: \"39e12872-2efe-4cec-82dd-a265f58da9ec\") " pod="openstack/dnsmasq-dns-594b65fc49-cgfb2" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.514052 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9j62\" (UniqueName: \"kubernetes.io/projected/39e12872-2efe-4cec-82dd-a265f58da9ec-kube-api-access-s9j62\") pod \"dnsmasq-dns-594b65fc49-cgfb2\" (UID: \"39e12872-2efe-4cec-82dd-a265f58da9ec\") " pod="openstack/dnsmasq-dns-594b65fc49-cgfb2" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.514420 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39e12872-2efe-4cec-82dd-a265f58da9ec-dns-svc\") pod \"dnsmasq-dns-594b65fc49-cgfb2\" (UID: \"39e12872-2efe-4cec-82dd-a265f58da9ec\") " pod="openstack/dnsmasq-dns-594b65fc49-cgfb2" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.514476 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39e12872-2efe-4cec-82dd-a265f58da9ec-config\") pod \"dnsmasq-dns-594b65fc49-cgfb2\" (UID: \"39e12872-2efe-4cec-82dd-a265f58da9ec\") " pod="openstack/dnsmasq-dns-594b65fc49-cgfb2" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.515451 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39e12872-2efe-4cec-82dd-a265f58da9ec-dns-svc\") pod \"dnsmasq-dns-594b65fc49-cgfb2\" (UID: \"39e12872-2efe-4cec-82dd-a265f58da9ec\") " pod="openstack/dnsmasq-dns-594b65fc49-cgfb2" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.515503 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39e12872-2efe-4cec-82dd-a265f58da9ec-config\") pod \"dnsmasq-dns-594b65fc49-cgfb2\" (UID: \"39e12872-2efe-4cec-82dd-a265f58da9ec\") " pod="openstack/dnsmasq-dns-594b65fc49-cgfb2" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.545597 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9j62\" (UniqueName: \"kubernetes.io/projected/39e12872-2efe-4cec-82dd-a265f58da9ec-kube-api-access-s9j62\") pod \"dnsmasq-dns-594b65fc49-cgfb2\" (UID: \"39e12872-2efe-4cec-82dd-a265f58da9ec\") " pod="openstack/dnsmasq-dns-594b65fc49-cgfb2" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.666183 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-594b65fc49-cgfb2" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.819785 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.821143 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.828241 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-74b2s" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.828451 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.828513 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.828569 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.828662 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.829014 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.830833 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 16:21:49 crc kubenswrapper[4998]: I1203 16:21:49.836591 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.020618 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/aeb4a971-3c03-457c-b0df-1503701aac5b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.020745 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/aeb4a971-3c03-457c-b0df-1503701aac5b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.020873 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/aeb4a971-3c03-457c-b0df-1503701aac5b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.020906 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/aeb4a971-3c03-457c-b0df-1503701aac5b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.021021 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aeb4a971-3c03-457c-b0df-1503701aac5b-config-data\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.021053 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/aeb4a971-3c03-457c-b0df-1503701aac5b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.021075 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/aeb4a971-3c03-457c-b0df-1503701aac5b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.021112 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/aeb4a971-3c03-457c-b0df-1503701aac5b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.021214 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdpq5\" (UniqueName: \"kubernetes.io/projected/aeb4a971-3c03-457c-b0df-1503701aac5b-kube-api-access-hdpq5\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.021265 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/aeb4a971-3c03-457c-b0df-1503701aac5b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.021287 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.101242 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.102358 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.105188 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.105321 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.105342 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.105680 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.105857 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-hpxvh" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.105971 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.106679 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.122292 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aeb4a971-3c03-457c-b0df-1503701aac5b-config-data\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.122349 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/aeb4a971-3c03-457c-b0df-1503701aac5b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.122386 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/aeb4a971-3c03-457c-b0df-1503701aac5b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.122442 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/aeb4a971-3c03-457c-b0df-1503701aac5b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.122485 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdpq5\" (UniqueName: \"kubernetes.io/projected/aeb4a971-3c03-457c-b0df-1503701aac5b-kube-api-access-hdpq5\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.122513 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/aeb4a971-3c03-457c-b0df-1503701aac5b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.122535 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.122561 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/aeb4a971-3c03-457c-b0df-1503701aac5b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.122594 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/aeb4a971-3c03-457c-b0df-1503701aac5b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.122606 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.122627 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/aeb4a971-3c03-457c-b0df-1503701aac5b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.122656 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/aeb4a971-3c03-457c-b0df-1503701aac5b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.123040 4998 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.123124 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aeb4a971-3c03-457c-b0df-1503701aac5b-config-data\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.123238 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/aeb4a971-3c03-457c-b0df-1503701aac5b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.123988 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/aeb4a971-3c03-457c-b0df-1503701aac5b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.124196 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/aeb4a971-3c03-457c-b0df-1503701aac5b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.135109 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/aeb4a971-3c03-457c-b0df-1503701aac5b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.135781 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/aeb4a971-3c03-457c-b0df-1503701aac5b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.136856 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/aeb4a971-3c03-457c-b0df-1503701aac5b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.139196 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/aeb4a971-3c03-457c-b0df-1503701aac5b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.157250 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/aeb4a971-3c03-457c-b0df-1503701aac5b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.160427 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdpq5\" (UniqueName: \"kubernetes.io/projected/aeb4a971-3c03-457c-b0df-1503701aac5b-kube-api-access-hdpq5\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.198887 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.223563 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e0ef6669-5a8b-462d-9681-925859c86df1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.223628 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e0ef6669-5a8b-462d-9681-925859c86df1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.223661 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bjlg\" (UniqueName: \"kubernetes.io/projected/e0ef6669-5a8b-462d-9681-925859c86df1-kube-api-access-7bjlg\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.223690 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.223727 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e0ef6669-5a8b-462d-9681-925859c86df1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.223746 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e0ef6669-5a8b-462d-9681-925859c86df1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.223785 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e0ef6669-5a8b-462d-9681-925859c86df1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.223827 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e0ef6669-5a8b-462d-9681-925859c86df1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.223858 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e0ef6669-5a8b-462d-9681-925859c86df1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.223896 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e0ef6669-5a8b-462d-9681-925859c86df1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.223921 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e0ef6669-5a8b-462d-9681-925859c86df1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.324707 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e0ef6669-5a8b-462d-9681-925859c86df1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.324798 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e0ef6669-5a8b-462d-9681-925859c86df1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.324822 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e0ef6669-5a8b-462d-9681-925859c86df1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.324863 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e0ef6669-5a8b-462d-9681-925859c86df1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.324888 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e0ef6669-5a8b-462d-9681-925859c86df1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.324915 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bjlg\" (UniqueName: \"kubernetes.io/projected/e0ef6669-5a8b-462d-9681-925859c86df1-kube-api-access-7bjlg\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.324942 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.324980 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e0ef6669-5a8b-462d-9681-925859c86df1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.325002 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e0ef6669-5a8b-462d-9681-925859c86df1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.325028 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e0ef6669-5a8b-462d-9681-925859c86df1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.325070 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e0ef6669-5a8b-462d-9681-925859c86df1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.326094 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e0ef6669-5a8b-462d-9681-925859c86df1-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.326188 4998 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.326535 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e0ef6669-5a8b-462d-9681-925859c86df1-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.326725 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e0ef6669-5a8b-462d-9681-925859c86df1-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.327176 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e0ef6669-5a8b-462d-9681-925859c86df1-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.331123 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e0ef6669-5a8b-462d-9681-925859c86df1-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.331341 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e0ef6669-5a8b-462d-9681-925859c86df1-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.343583 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e0ef6669-5a8b-462d-9681-925859c86df1-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.351492 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e0ef6669-5a8b-462d-9681-925859c86df1-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.353790 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bjlg\" (UniqueName: \"kubernetes.io/projected/e0ef6669-5a8b-462d-9681-925859c86df1-kube-api-access-7bjlg\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.354316 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e0ef6669-5a8b-462d-9681-925859c86df1-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.360960 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.418308 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.439199 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.503655 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-notifications-server-0"] Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.505041 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.507863 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-notifications-config-data" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.508111 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-notifications-default-user" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.508289 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-notifications-server-dockercfg-dxj9s" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.508486 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-notifications-erlang-cookie" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.508631 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-notifications-svc" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.508736 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-notifications-plugins-conf" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.521449 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-notifications-server-conf" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.537104 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-notifications-server-0"] Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.629619 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dbde2166-b545-4979-845a-8932845b9cd1-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.629784 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dbde2166-b545-4979-845a-8932845b9cd1-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.629878 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dbde2166-b545-4979-845a-8932845b9cd1-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.629983 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.630050 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dbde2166-b545-4979-845a-8932845b9cd1-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.630068 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dbde2166-b545-4979-845a-8932845b9cd1-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.630229 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dbde2166-b545-4979-845a-8932845b9cd1-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.630309 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dbde2166-b545-4979-845a-8932845b9cd1-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.630387 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dbde2166-b545-4979-845a-8932845b9cd1-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.630469 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpjjg\" (UniqueName: \"kubernetes.io/projected/dbde2166-b545-4979-845a-8932845b9cd1-kube-api-access-bpjjg\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.630523 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dbde2166-b545-4979-845a-8932845b9cd1-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.732477 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpjjg\" (UniqueName: \"kubernetes.io/projected/dbde2166-b545-4979-845a-8932845b9cd1-kube-api-access-bpjjg\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.732910 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dbde2166-b545-4979-845a-8932845b9cd1-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.733558 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dbde2166-b545-4979-845a-8932845b9cd1-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.733601 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dbde2166-b545-4979-845a-8932845b9cd1-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.733631 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dbde2166-b545-4979-845a-8932845b9cd1-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.733671 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.733702 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dbde2166-b545-4979-845a-8932845b9cd1-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.733723 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dbde2166-b545-4979-845a-8932845b9cd1-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.733779 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dbde2166-b545-4979-845a-8932845b9cd1-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.733811 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dbde2166-b545-4979-845a-8932845b9cd1-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.733938 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dbde2166-b545-4979-845a-8932845b9cd1-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.733990 4998 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.734102 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dbde2166-b545-4979-845a-8932845b9cd1-rabbitmq-erlang-cookie\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.734474 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dbde2166-b545-4979-845a-8932845b9cd1-rabbitmq-plugins\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.734623 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dbde2166-b545-4979-845a-8932845b9cd1-config-data\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.734837 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dbde2166-b545-4979-845a-8932845b9cd1-server-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.735094 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dbde2166-b545-4979-845a-8932845b9cd1-plugins-conf\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.737435 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dbde2166-b545-4979-845a-8932845b9cd1-erlang-cookie-secret\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.738132 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dbde2166-b545-4979-845a-8932845b9cd1-rabbitmq-tls\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.738412 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dbde2166-b545-4979-845a-8932845b9cd1-rabbitmq-confd\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.739004 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dbde2166-b545-4979-845a-8932845b9cd1-pod-info\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.748362 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpjjg\" (UniqueName: \"kubernetes.io/projected/dbde2166-b545-4979-845a-8932845b9cd1-kube-api-access-bpjjg\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.763895 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-notifications-server-0\" (UID: \"dbde2166-b545-4979-845a-8932845b9cd1\") " pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:50 crc kubenswrapper[4998]: I1203 16:21:50.831880 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.504080 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.505724 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.508913 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-tknt6" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.511924 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.513767 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.513942 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.515020 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.521541 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.665440 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ea3f8b4-86fc-4888-b9c0-36859ab25724-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"8ea3f8b4-86fc-4888-b9c0-36859ab25724\") " pod="openstack/openstack-galera-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.665511 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"8ea3f8b4-86fc-4888-b9c0-36859ab25724\") " pod="openstack/openstack-galera-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.665585 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ea3f8b4-86fc-4888-b9c0-36859ab25724-operator-scripts\") pod \"openstack-galera-0\" (UID: \"8ea3f8b4-86fc-4888-b9c0-36859ab25724\") " pod="openstack/openstack-galera-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.665618 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8ea3f8b4-86fc-4888-b9c0-36859ab25724-kolla-config\") pod \"openstack-galera-0\" (UID: \"8ea3f8b4-86fc-4888-b9c0-36859ab25724\") " pod="openstack/openstack-galera-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.665639 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pz62\" (UniqueName: \"kubernetes.io/projected/8ea3f8b4-86fc-4888-b9c0-36859ab25724-kube-api-access-8pz62\") pod \"openstack-galera-0\" (UID: \"8ea3f8b4-86fc-4888-b9c0-36859ab25724\") " pod="openstack/openstack-galera-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.665666 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ea3f8b4-86fc-4888-b9c0-36859ab25724-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"8ea3f8b4-86fc-4888-b9c0-36859ab25724\") " pod="openstack/openstack-galera-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.665698 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8ea3f8b4-86fc-4888-b9c0-36859ab25724-config-data-default\") pod \"openstack-galera-0\" (UID: \"8ea3f8b4-86fc-4888-b9c0-36859ab25724\") " pod="openstack/openstack-galera-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.666106 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8ea3f8b4-86fc-4888-b9c0-36859ab25724-config-data-generated\") pod \"openstack-galera-0\" (UID: \"8ea3f8b4-86fc-4888-b9c0-36859ab25724\") " pod="openstack/openstack-galera-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.766982 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ea3f8b4-86fc-4888-b9c0-36859ab25724-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"8ea3f8b4-86fc-4888-b9c0-36859ab25724\") " pod="openstack/openstack-galera-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.767046 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"8ea3f8b4-86fc-4888-b9c0-36859ab25724\") " pod="openstack/openstack-galera-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.767104 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ea3f8b4-86fc-4888-b9c0-36859ab25724-operator-scripts\") pod \"openstack-galera-0\" (UID: \"8ea3f8b4-86fc-4888-b9c0-36859ab25724\") " pod="openstack/openstack-galera-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.767125 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8ea3f8b4-86fc-4888-b9c0-36859ab25724-kolla-config\") pod \"openstack-galera-0\" (UID: \"8ea3f8b4-86fc-4888-b9c0-36859ab25724\") " pod="openstack/openstack-galera-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.767142 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pz62\" (UniqueName: \"kubernetes.io/projected/8ea3f8b4-86fc-4888-b9c0-36859ab25724-kube-api-access-8pz62\") pod \"openstack-galera-0\" (UID: \"8ea3f8b4-86fc-4888-b9c0-36859ab25724\") " pod="openstack/openstack-galera-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.767162 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ea3f8b4-86fc-4888-b9c0-36859ab25724-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"8ea3f8b4-86fc-4888-b9c0-36859ab25724\") " pod="openstack/openstack-galera-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.767200 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8ea3f8b4-86fc-4888-b9c0-36859ab25724-config-data-default\") pod \"openstack-galera-0\" (UID: \"8ea3f8b4-86fc-4888-b9c0-36859ab25724\") " pod="openstack/openstack-galera-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.767219 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8ea3f8b4-86fc-4888-b9c0-36859ab25724-config-data-generated\") pod \"openstack-galera-0\" (UID: \"8ea3f8b4-86fc-4888-b9c0-36859ab25724\") " pod="openstack/openstack-galera-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.767592 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8ea3f8b4-86fc-4888-b9c0-36859ab25724-config-data-generated\") pod \"openstack-galera-0\" (UID: \"8ea3f8b4-86fc-4888-b9c0-36859ab25724\") " pod="openstack/openstack-galera-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.767895 4998 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"8ea3f8b4-86fc-4888-b9c0-36859ab25724\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-galera-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.769039 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8ea3f8b4-86fc-4888-b9c0-36859ab25724-kolla-config\") pod \"openstack-galera-0\" (UID: \"8ea3f8b4-86fc-4888-b9c0-36859ab25724\") " pod="openstack/openstack-galera-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.769389 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8ea3f8b4-86fc-4888-b9c0-36859ab25724-config-data-default\") pod \"openstack-galera-0\" (UID: \"8ea3f8b4-86fc-4888-b9c0-36859ab25724\") " pod="openstack/openstack-galera-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.769982 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ea3f8b4-86fc-4888-b9c0-36859ab25724-operator-scripts\") pod \"openstack-galera-0\" (UID: \"8ea3f8b4-86fc-4888-b9c0-36859ab25724\") " pod="openstack/openstack-galera-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.772627 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ea3f8b4-86fc-4888-b9c0-36859ab25724-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"8ea3f8b4-86fc-4888-b9c0-36859ab25724\") " pod="openstack/openstack-galera-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.789425 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ea3f8b4-86fc-4888-b9c0-36859ab25724-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"8ea3f8b4-86fc-4888-b9c0-36859ab25724\") " pod="openstack/openstack-galera-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.824193 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"8ea3f8b4-86fc-4888-b9c0-36859ab25724\") " pod="openstack/openstack-galera-0" Dec 03 16:21:52 crc kubenswrapper[4998]: I1203 16:21:52.824614 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pz62\" (UniqueName: \"kubernetes.io/projected/8ea3f8b4-86fc-4888-b9c0-36859ab25724-kube-api-access-8pz62\") pod \"openstack-galera-0\" (UID: \"8ea3f8b4-86fc-4888-b9c0-36859ab25724\") " pod="openstack/openstack-galera-0" Dec 03 16:21:53 crc kubenswrapper[4998]: I1203 16:21:53.126892 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.097212 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.098557 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.101080 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-hdm9q" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.101135 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.101386 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.101586 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.109985 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.156119 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.157523 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.161749 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.162120 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-h4kdb" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.162118 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.188682 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.192566 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b2c3df0-703c-438e-9887-cf29844b9eb7-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1b2c3df0-703c-438e-9887-cf29844b9eb7\") " pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.192609 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1b2c3df0-703c-438e-9887-cf29844b9eb7-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1b2c3df0-703c-438e-9887-cf29844b9eb7\") " pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.192642 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2c3df0-703c-438e-9887-cf29844b9eb7-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1b2c3df0-703c-438e-9887-cf29844b9eb7\") " pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.192662 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1b2c3df0-703c-438e-9887-cf29844b9eb7-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1b2c3df0-703c-438e-9887-cf29844b9eb7\") " pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.192698 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1b2c3df0-703c-438e-9887-cf29844b9eb7-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1b2c3df0-703c-438e-9887-cf29844b9eb7\") " pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.192832 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1b2c3df0-703c-438e-9887-cf29844b9eb7\") " pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.192853 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pxtt\" (UniqueName: \"kubernetes.io/projected/1b2c3df0-703c-438e-9887-cf29844b9eb7-kube-api-access-8pxtt\") pod \"openstack-cell1-galera-0\" (UID: \"1b2c3df0-703c-438e-9887-cf29844b9eb7\") " pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.192872 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1b2c3df0-703c-438e-9887-cf29844b9eb7-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1b2c3df0-703c-438e-9887-cf29844b9eb7\") " pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.294164 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1b2c3df0-703c-438e-9887-cf29844b9eb7-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1b2c3df0-703c-438e-9887-cf29844b9eb7\") " pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.294222 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1b2c3df0-703c-438e-9887-cf29844b9eb7-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1b2c3df0-703c-438e-9887-cf29844b9eb7\") " pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.294262 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/75024598-618b-49e3-919c-3bf92e4251a2-config-data\") pod \"memcached-0\" (UID: \"75024598-618b-49e3-919c-3bf92e4251a2\") " pod="openstack/memcached-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.294278 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xzrh\" (UniqueName: \"kubernetes.io/projected/75024598-618b-49e3-919c-3bf92e4251a2-kube-api-access-4xzrh\") pod \"memcached-0\" (UID: \"75024598-618b-49e3-919c-3bf92e4251a2\") " pod="openstack/memcached-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.294300 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1b2c3df0-703c-438e-9887-cf29844b9eb7\") " pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.294318 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pxtt\" (UniqueName: \"kubernetes.io/projected/1b2c3df0-703c-438e-9887-cf29844b9eb7-kube-api-access-8pxtt\") pod \"openstack-cell1-galera-0\" (UID: \"1b2c3df0-703c-438e-9887-cf29844b9eb7\") " pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.294338 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1b2c3df0-703c-438e-9887-cf29844b9eb7-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1b2c3df0-703c-438e-9887-cf29844b9eb7\") " pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.294370 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75024598-618b-49e3-919c-3bf92e4251a2-combined-ca-bundle\") pod \"memcached-0\" (UID: \"75024598-618b-49e3-919c-3bf92e4251a2\") " pod="openstack/memcached-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.294393 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/75024598-618b-49e3-919c-3bf92e4251a2-kolla-config\") pod \"memcached-0\" (UID: \"75024598-618b-49e3-919c-3bf92e4251a2\") " pod="openstack/memcached-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.294433 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b2c3df0-703c-438e-9887-cf29844b9eb7-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1b2c3df0-703c-438e-9887-cf29844b9eb7\") " pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.294451 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1b2c3df0-703c-438e-9887-cf29844b9eb7-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1b2c3df0-703c-438e-9887-cf29844b9eb7\") " pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.294473 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/75024598-618b-49e3-919c-3bf92e4251a2-memcached-tls-certs\") pod \"memcached-0\" (UID: \"75024598-618b-49e3-919c-3bf92e4251a2\") " pod="openstack/memcached-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.294495 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2c3df0-703c-438e-9887-cf29844b9eb7-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1b2c3df0-703c-438e-9887-cf29844b9eb7\") " pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.295836 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1b2c3df0-703c-438e-9887-cf29844b9eb7-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1b2c3df0-703c-438e-9887-cf29844b9eb7\") " pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.295911 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1b2c3df0-703c-438e-9887-cf29844b9eb7-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1b2c3df0-703c-438e-9887-cf29844b9eb7\") " pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.296118 4998 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1b2c3df0-703c-438e-9887-cf29844b9eb7\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.296247 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1b2c3df0-703c-438e-9887-cf29844b9eb7-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1b2c3df0-703c-438e-9887-cf29844b9eb7\") " pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.296165 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1b2c3df0-703c-438e-9887-cf29844b9eb7-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1b2c3df0-703c-438e-9887-cf29844b9eb7\") " pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.299795 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2c3df0-703c-438e-9887-cf29844b9eb7-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1b2c3df0-703c-438e-9887-cf29844b9eb7\") " pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.301065 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b2c3df0-703c-438e-9887-cf29844b9eb7-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1b2c3df0-703c-438e-9887-cf29844b9eb7\") " pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.328180 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1b2c3df0-703c-438e-9887-cf29844b9eb7\") " pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.332472 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pxtt\" (UniqueName: \"kubernetes.io/projected/1b2c3df0-703c-438e-9887-cf29844b9eb7-kube-api-access-8pxtt\") pod \"openstack-cell1-galera-0\" (UID: \"1b2c3df0-703c-438e-9887-cf29844b9eb7\") " pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.395831 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/75024598-618b-49e3-919c-3bf92e4251a2-memcached-tls-certs\") pod \"memcached-0\" (UID: \"75024598-618b-49e3-919c-3bf92e4251a2\") " pod="openstack/memcached-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.395919 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xzrh\" (UniqueName: \"kubernetes.io/projected/75024598-618b-49e3-919c-3bf92e4251a2-kube-api-access-4xzrh\") pod \"memcached-0\" (UID: \"75024598-618b-49e3-919c-3bf92e4251a2\") " pod="openstack/memcached-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.395939 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/75024598-618b-49e3-919c-3bf92e4251a2-config-data\") pod \"memcached-0\" (UID: \"75024598-618b-49e3-919c-3bf92e4251a2\") " pod="openstack/memcached-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.395986 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75024598-618b-49e3-919c-3bf92e4251a2-combined-ca-bundle\") pod \"memcached-0\" (UID: \"75024598-618b-49e3-919c-3bf92e4251a2\") " pod="openstack/memcached-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.396017 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/75024598-618b-49e3-919c-3bf92e4251a2-kolla-config\") pod \"memcached-0\" (UID: \"75024598-618b-49e3-919c-3bf92e4251a2\") " pod="openstack/memcached-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.397225 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/75024598-618b-49e3-919c-3bf92e4251a2-kolla-config\") pod \"memcached-0\" (UID: \"75024598-618b-49e3-919c-3bf92e4251a2\") " pod="openstack/memcached-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.397241 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/75024598-618b-49e3-919c-3bf92e4251a2-config-data\") pod \"memcached-0\" (UID: \"75024598-618b-49e3-919c-3bf92e4251a2\") " pod="openstack/memcached-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.401342 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75024598-618b-49e3-919c-3bf92e4251a2-combined-ca-bundle\") pod \"memcached-0\" (UID: \"75024598-618b-49e3-919c-3bf92e4251a2\") " pod="openstack/memcached-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.401368 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/75024598-618b-49e3-919c-3bf92e4251a2-memcached-tls-certs\") pod \"memcached-0\" (UID: \"75024598-618b-49e3-919c-3bf92e4251a2\") " pod="openstack/memcached-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.419380 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xzrh\" (UniqueName: \"kubernetes.io/projected/75024598-618b-49e3-919c-3bf92e4251a2-kube-api-access-4xzrh\") pod \"memcached-0\" (UID: \"75024598-618b-49e3-919c-3bf92e4251a2\") " pod="openstack/memcached-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.419734 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 16:21:54 crc kubenswrapper[4998]: I1203 16:21:54.482499 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 16:21:56 crc kubenswrapper[4998]: I1203 16:21:56.292592 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 16:21:56 crc kubenswrapper[4998]: I1203 16:21:56.308806 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 16:21:56 crc kubenswrapper[4998]: I1203 16:21:56.312245 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-5rhqp" Dec 03 16:21:56 crc kubenswrapper[4998]: I1203 16:21:56.321903 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 16:21:56 crc kubenswrapper[4998]: I1203 16:21:56.435698 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xz9qq\" (UniqueName: \"kubernetes.io/projected/f9abb1de-7121-4de9-833a-9f1ab889b8ff-kube-api-access-xz9qq\") pod \"kube-state-metrics-0\" (UID: \"f9abb1de-7121-4de9-833a-9f1ab889b8ff\") " pod="openstack/kube-state-metrics-0" Dec 03 16:21:56 crc kubenswrapper[4998]: I1203 16:21:56.537681 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xz9qq\" (UniqueName: \"kubernetes.io/projected/f9abb1de-7121-4de9-833a-9f1ab889b8ff-kube-api-access-xz9qq\") pod \"kube-state-metrics-0\" (UID: \"f9abb1de-7121-4de9-833a-9f1ab889b8ff\") " pod="openstack/kube-state-metrics-0" Dec 03 16:21:56 crc kubenswrapper[4998]: I1203 16:21:56.557677 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xz9qq\" (UniqueName: \"kubernetes.io/projected/f9abb1de-7121-4de9-833a-9f1ab889b8ff-kube-api-access-xz9qq\") pod \"kube-state-metrics-0\" (UID: \"f9abb1de-7121-4de9-833a-9f1ab889b8ff\") " pod="openstack/kube-state-metrics-0" Dec 03 16:21:56 crc kubenswrapper[4998]: I1203 16:21:56.635461 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.111897 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.112003 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.520691 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.523811 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.526499 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.526901 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.526948 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.528009 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-fx7rj" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.532493 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.537956 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.544144 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.655534 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.655606 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c9889e3e-2478-469f-b92d-e203678c4927\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c9889e3e-2478-469f-b92d-e203678c4927\") pod \"prometheus-metric-storage-0\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.655638 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.655662 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.655698 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz4xj\" (UniqueName: \"kubernetes.io/projected/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-kube-api-access-rz4xj\") pod \"prometheus-metric-storage-0\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.655721 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.655741 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.655779 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-config\") pod \"prometheus-metric-storage-0\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.757082 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.757178 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz4xj\" (UniqueName: \"kubernetes.io/projected/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-kube-api-access-rz4xj\") pod \"prometheus-metric-storage-0\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.757223 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.757261 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.757307 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-config\") pod \"prometheus-metric-storage-0\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.757429 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.757466 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c9889e3e-2478-469f-b92d-e203678c4927\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c9889e3e-2478-469f-b92d-e203678c4927\") pod \"prometheus-metric-storage-0\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.757528 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.761978 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.767427 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.769445 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-config\") pod \"prometheus-metric-storage-0\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.771649 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.774357 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.775561 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.776444 4998 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.776516 4998 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c9889e3e-2478-469f-b92d-e203678c4927\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c9889e3e-2478-469f-b92d-e203678c4927\") pod \"prometheus-metric-storage-0\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/afadf41a682809fbeb8f0a8dc0e25f598b7474e7bc0a2ee0471a59f3f2d1c942/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.790385 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz4xj\" (UniqueName: \"kubernetes.io/projected/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-kube-api-access-rz4xj\") pod \"prometheus-metric-storage-0\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:57 crc kubenswrapper[4998]: I1203 16:21:57.973161 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c9889e3e-2478-469f-b92d-e203678c4927\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c9889e3e-2478-469f-b92d-e203678c4927\") pod \"prometheus-metric-storage-0\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:58 crc kubenswrapper[4998]: I1203 16:21:58.149820 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.668197 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-7sxbw"] Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.669931 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7sxbw" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.672268 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-6hbjd" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.672477 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.672474 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.688257 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-7sxbw"] Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.696125 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-5hxf2"] Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.698482 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-5hxf2" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.723064 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-5hxf2"] Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.792842 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rnmq\" (UniqueName: \"kubernetes.io/projected/832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5-kube-api-access-7rnmq\") pod \"ovn-controller-ovs-5hxf2\" (UID: \"832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5\") " pod="openstack/ovn-controller-ovs-5hxf2" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.792905 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f-combined-ca-bundle\") pod \"ovn-controller-7sxbw\" (UID: \"95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f\") " pod="openstack/ovn-controller-7sxbw" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.793050 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f-ovn-controller-tls-certs\") pod \"ovn-controller-7sxbw\" (UID: \"95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f\") " pod="openstack/ovn-controller-7sxbw" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.793102 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5-var-lib\") pod \"ovn-controller-ovs-5hxf2\" (UID: \"832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5\") " pod="openstack/ovn-controller-ovs-5hxf2" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.793119 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5-etc-ovs\") pod \"ovn-controller-ovs-5hxf2\" (UID: \"832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5\") " pod="openstack/ovn-controller-ovs-5hxf2" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.793138 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5-scripts\") pod \"ovn-controller-ovs-5hxf2\" (UID: \"832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5\") " pod="openstack/ovn-controller-ovs-5hxf2" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.793191 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5-var-run\") pod \"ovn-controller-ovs-5hxf2\" (UID: \"832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5\") " pod="openstack/ovn-controller-ovs-5hxf2" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.793244 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qngd2\" (UniqueName: \"kubernetes.io/projected/95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f-kube-api-access-qngd2\") pod \"ovn-controller-7sxbw\" (UID: \"95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f\") " pod="openstack/ovn-controller-7sxbw" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.793280 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f-var-run\") pod \"ovn-controller-7sxbw\" (UID: \"95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f\") " pod="openstack/ovn-controller-7sxbw" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.793303 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5-var-log\") pod \"ovn-controller-ovs-5hxf2\" (UID: \"832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5\") " pod="openstack/ovn-controller-ovs-5hxf2" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.793353 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f-var-run-ovn\") pod \"ovn-controller-7sxbw\" (UID: \"95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f\") " pod="openstack/ovn-controller-7sxbw" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.793382 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f-var-log-ovn\") pod \"ovn-controller-7sxbw\" (UID: \"95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f\") " pod="openstack/ovn-controller-7sxbw" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.793397 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f-scripts\") pod \"ovn-controller-7sxbw\" (UID: \"95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f\") " pod="openstack/ovn-controller-7sxbw" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.895694 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rnmq\" (UniqueName: \"kubernetes.io/projected/832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5-kube-api-access-7rnmq\") pod \"ovn-controller-ovs-5hxf2\" (UID: \"832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5\") " pod="openstack/ovn-controller-ovs-5hxf2" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.895748 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f-combined-ca-bundle\") pod \"ovn-controller-7sxbw\" (UID: \"95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f\") " pod="openstack/ovn-controller-7sxbw" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.895806 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f-ovn-controller-tls-certs\") pod \"ovn-controller-7sxbw\" (UID: \"95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f\") " pod="openstack/ovn-controller-7sxbw" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.895839 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5-var-lib\") pod \"ovn-controller-ovs-5hxf2\" (UID: \"832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5\") " pod="openstack/ovn-controller-ovs-5hxf2" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.895861 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5-etc-ovs\") pod \"ovn-controller-ovs-5hxf2\" (UID: \"832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5\") " pod="openstack/ovn-controller-ovs-5hxf2" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.895886 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5-scripts\") pod \"ovn-controller-ovs-5hxf2\" (UID: \"832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5\") " pod="openstack/ovn-controller-ovs-5hxf2" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.895924 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5-var-run\") pod \"ovn-controller-ovs-5hxf2\" (UID: \"832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5\") " pod="openstack/ovn-controller-ovs-5hxf2" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.895965 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qngd2\" (UniqueName: \"kubernetes.io/projected/95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f-kube-api-access-qngd2\") pod \"ovn-controller-7sxbw\" (UID: \"95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f\") " pod="openstack/ovn-controller-7sxbw" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.895996 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f-var-run\") pod \"ovn-controller-7sxbw\" (UID: \"95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f\") " pod="openstack/ovn-controller-7sxbw" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.896021 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5-var-log\") pod \"ovn-controller-ovs-5hxf2\" (UID: \"832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5\") " pod="openstack/ovn-controller-ovs-5hxf2" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.896059 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f-var-run-ovn\") pod \"ovn-controller-7sxbw\" (UID: \"95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f\") " pod="openstack/ovn-controller-7sxbw" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.896085 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f-var-log-ovn\") pod \"ovn-controller-7sxbw\" (UID: \"95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f\") " pod="openstack/ovn-controller-7sxbw" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.896104 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f-scripts\") pod \"ovn-controller-7sxbw\" (UID: \"95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f\") " pod="openstack/ovn-controller-7sxbw" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.897165 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5-var-run\") pod \"ovn-controller-ovs-5hxf2\" (UID: \"832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5\") " pod="openstack/ovn-controller-ovs-5hxf2" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.897201 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f-var-log-ovn\") pod \"ovn-controller-7sxbw\" (UID: \"95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f\") " pod="openstack/ovn-controller-7sxbw" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.897260 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5-var-log\") pod \"ovn-controller-ovs-5hxf2\" (UID: \"832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5\") " pod="openstack/ovn-controller-ovs-5hxf2" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.897267 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f-var-run-ovn\") pod \"ovn-controller-7sxbw\" (UID: \"95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f\") " pod="openstack/ovn-controller-7sxbw" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.897265 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f-var-run\") pod \"ovn-controller-7sxbw\" (UID: \"95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f\") " pod="openstack/ovn-controller-7sxbw" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.897464 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5-etc-ovs\") pod \"ovn-controller-ovs-5hxf2\" (UID: \"832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5\") " pod="openstack/ovn-controller-ovs-5hxf2" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.897468 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5-var-lib\") pod \"ovn-controller-ovs-5hxf2\" (UID: \"832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5\") " pod="openstack/ovn-controller-ovs-5hxf2" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.899084 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f-scripts\") pod \"ovn-controller-7sxbw\" (UID: \"95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f\") " pod="openstack/ovn-controller-7sxbw" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.899897 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5-scripts\") pod \"ovn-controller-ovs-5hxf2\" (UID: \"832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5\") " pod="openstack/ovn-controller-ovs-5hxf2" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.904792 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f-combined-ca-bundle\") pod \"ovn-controller-7sxbw\" (UID: \"95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f\") " pod="openstack/ovn-controller-7sxbw" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.916432 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f-ovn-controller-tls-certs\") pod \"ovn-controller-7sxbw\" (UID: \"95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f\") " pod="openstack/ovn-controller-7sxbw" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.929464 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qngd2\" (UniqueName: \"kubernetes.io/projected/95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f-kube-api-access-qngd2\") pod \"ovn-controller-7sxbw\" (UID: \"95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f\") " pod="openstack/ovn-controller-7sxbw" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.938522 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rnmq\" (UniqueName: \"kubernetes.io/projected/832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5-kube-api-access-7rnmq\") pod \"ovn-controller-ovs-5hxf2\" (UID: \"832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5\") " pod="openstack/ovn-controller-ovs-5hxf2" Dec 03 16:21:59 crc kubenswrapper[4998]: I1203 16:21:59.997845 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7sxbw" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.022412 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-5hxf2" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.146481 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.150117 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.151659 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.152460 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-llhrz" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.152658 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.152854 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.153066 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.160236 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.199851 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/893cf12f-77a9-463f-8dd4-1b3f05b2c9bf-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf\") " pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.199979 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/893cf12f-77a9-463f-8dd4-1b3f05b2c9bf-config\") pod \"ovsdbserver-nb-0\" (UID: \"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf\") " pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.200025 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/893cf12f-77a9-463f-8dd4-1b3f05b2c9bf-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf\") " pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.200052 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/893cf12f-77a9-463f-8dd4-1b3f05b2c9bf-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf\") " pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.200083 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/893cf12f-77a9-463f-8dd4-1b3f05b2c9bf-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf\") " pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.200112 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf\") " pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.200200 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/893cf12f-77a9-463f-8dd4-1b3f05b2c9bf-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf\") " pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.200395 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmmsq\" (UniqueName: \"kubernetes.io/projected/893cf12f-77a9-463f-8dd4-1b3f05b2c9bf-kube-api-access-dmmsq\") pod \"ovsdbserver-nb-0\" (UID: \"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf\") " pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.302406 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmmsq\" (UniqueName: \"kubernetes.io/projected/893cf12f-77a9-463f-8dd4-1b3f05b2c9bf-kube-api-access-dmmsq\") pod \"ovsdbserver-nb-0\" (UID: \"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf\") " pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.302470 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/893cf12f-77a9-463f-8dd4-1b3f05b2c9bf-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf\") " pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.302527 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/893cf12f-77a9-463f-8dd4-1b3f05b2c9bf-config\") pod \"ovsdbserver-nb-0\" (UID: \"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf\") " pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.302554 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/893cf12f-77a9-463f-8dd4-1b3f05b2c9bf-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf\") " pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.302582 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/893cf12f-77a9-463f-8dd4-1b3f05b2c9bf-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf\") " pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.302615 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/893cf12f-77a9-463f-8dd4-1b3f05b2c9bf-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf\") " pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.302648 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf\") " pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.302671 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/893cf12f-77a9-463f-8dd4-1b3f05b2c9bf-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf\") " pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.303242 4998 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.303435 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/893cf12f-77a9-463f-8dd4-1b3f05b2c9bf-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf\") " pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.303679 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/893cf12f-77a9-463f-8dd4-1b3f05b2c9bf-config\") pod \"ovsdbserver-nb-0\" (UID: \"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf\") " pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.304590 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/893cf12f-77a9-463f-8dd4-1b3f05b2c9bf-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf\") " pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.307576 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/893cf12f-77a9-463f-8dd4-1b3f05b2c9bf-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf\") " pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.308425 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/893cf12f-77a9-463f-8dd4-1b3f05b2c9bf-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf\") " pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.319556 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/893cf12f-77a9-463f-8dd4-1b3f05b2c9bf-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf\") " pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.320324 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmmsq\" (UniqueName: \"kubernetes.io/projected/893cf12f-77a9-463f-8dd4-1b3f05b2c9bf-kube-api-access-dmmsq\") pod \"ovsdbserver-nb-0\" (UID: \"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf\") " pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.326730 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf\") " pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:00 crc kubenswrapper[4998]: I1203 16:22:00.470953 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.518130 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.521237 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.524613 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.524700 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.524740 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.524913 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-pkxvb" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.538222 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.664529 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37aa6aee-15f8-4e23-a18c-216b85e605cf-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"37aa6aee-15f8-4e23-a18c-216b85e605cf\") " pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.664579 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/37aa6aee-15f8-4e23-a18c-216b85e605cf-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"37aa6aee-15f8-4e23-a18c-216b85e605cf\") " pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.664633 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/37aa6aee-15f8-4e23-a18c-216b85e605cf-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"37aa6aee-15f8-4e23-a18c-216b85e605cf\") " pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.664655 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37aa6aee-15f8-4e23-a18c-216b85e605cf-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"37aa6aee-15f8-4e23-a18c-216b85e605cf\") " pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.664699 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"37aa6aee-15f8-4e23-a18c-216b85e605cf\") " pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.664719 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wn59\" (UniqueName: \"kubernetes.io/projected/37aa6aee-15f8-4e23-a18c-216b85e605cf-kube-api-access-2wn59\") pod \"ovsdbserver-sb-0\" (UID: \"37aa6aee-15f8-4e23-a18c-216b85e605cf\") " pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.664777 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37aa6aee-15f8-4e23-a18c-216b85e605cf-config\") pod \"ovsdbserver-sb-0\" (UID: \"37aa6aee-15f8-4e23-a18c-216b85e605cf\") " pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.664828 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/37aa6aee-15f8-4e23-a18c-216b85e605cf-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"37aa6aee-15f8-4e23-a18c-216b85e605cf\") " pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.766239 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37aa6aee-15f8-4e23-a18c-216b85e605cf-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"37aa6aee-15f8-4e23-a18c-216b85e605cf\") " pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.766296 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/37aa6aee-15f8-4e23-a18c-216b85e605cf-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"37aa6aee-15f8-4e23-a18c-216b85e605cf\") " pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.766381 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/37aa6aee-15f8-4e23-a18c-216b85e605cf-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"37aa6aee-15f8-4e23-a18c-216b85e605cf\") " pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.766410 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37aa6aee-15f8-4e23-a18c-216b85e605cf-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"37aa6aee-15f8-4e23-a18c-216b85e605cf\") " pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.766475 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"37aa6aee-15f8-4e23-a18c-216b85e605cf\") " pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.766505 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wn59\" (UniqueName: \"kubernetes.io/projected/37aa6aee-15f8-4e23-a18c-216b85e605cf-kube-api-access-2wn59\") pod \"ovsdbserver-sb-0\" (UID: \"37aa6aee-15f8-4e23-a18c-216b85e605cf\") " pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.766562 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37aa6aee-15f8-4e23-a18c-216b85e605cf-config\") pod \"ovsdbserver-sb-0\" (UID: \"37aa6aee-15f8-4e23-a18c-216b85e605cf\") " pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.766625 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/37aa6aee-15f8-4e23-a18c-216b85e605cf-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"37aa6aee-15f8-4e23-a18c-216b85e605cf\") " pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.767021 4998 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"37aa6aee-15f8-4e23-a18c-216b85e605cf\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.767333 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/37aa6aee-15f8-4e23-a18c-216b85e605cf-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"37aa6aee-15f8-4e23-a18c-216b85e605cf\") " pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.768242 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37aa6aee-15f8-4e23-a18c-216b85e605cf-config\") pod \"ovsdbserver-sb-0\" (UID: \"37aa6aee-15f8-4e23-a18c-216b85e605cf\") " pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.768890 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37aa6aee-15f8-4e23-a18c-216b85e605cf-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"37aa6aee-15f8-4e23-a18c-216b85e605cf\") " pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.773717 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37aa6aee-15f8-4e23-a18c-216b85e605cf-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"37aa6aee-15f8-4e23-a18c-216b85e605cf\") " pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.774502 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/37aa6aee-15f8-4e23-a18c-216b85e605cf-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"37aa6aee-15f8-4e23-a18c-216b85e605cf\") " pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.776137 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/37aa6aee-15f8-4e23-a18c-216b85e605cf-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"37aa6aee-15f8-4e23-a18c-216b85e605cf\") " pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.789366 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wn59\" (UniqueName: \"kubernetes.io/projected/37aa6aee-15f8-4e23-a18c-216b85e605cf-kube-api-access-2wn59\") pod \"ovsdbserver-sb-0\" (UID: \"37aa6aee-15f8-4e23-a18c-216b85e605cf\") " pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.813564 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"37aa6aee-15f8-4e23-a18c-216b85e605cf\") " pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:03 crc kubenswrapper[4998]: I1203 16:22:03.882287 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:06 crc kubenswrapper[4998]: I1203 16:22:06.027624 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7786d8fd7-fmq4c"] Dec 03 16:22:10 crc kubenswrapper[4998]: W1203 16:22:10.226344 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27c5a979_99f8_46a9_9a96_312c8c61186f.slice/crio-3fc343c2c1818f5d350875851d70b58b6ff0858d74aed9192c757679e6d36d42 WatchSource:0}: Error finding container 3fc343c2c1818f5d350875851d70b58b6ff0858d74aed9192c757679e6d36d42: Status 404 returned error can't find the container with id 3fc343c2c1818f5d350875851d70b58b6ff0858d74aed9192c757679e6d36d42 Dec 03 16:22:10 crc kubenswrapper[4998]: I1203 16:22:10.683789 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 16:22:10 crc kubenswrapper[4998]: I1203 16:22:10.745620 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 16:22:11 crc kubenswrapper[4998]: W1203 16:22:11.060599 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaeb4a971_3c03_457c_b0df_1503701aac5b.slice/crio-4b6f937511621fe03bbf3a85fbbb4a9d458cac75832cdb13200fa423199deea7 WatchSource:0}: Error finding container 4b6f937511621fe03bbf3a85fbbb4a9d458cac75832cdb13200fa423199deea7: Status 404 returned error can't find the container with id 4b6f937511621fe03bbf3a85fbbb4a9d458cac75832cdb13200fa423199deea7 Dec 03 16:22:11 crc kubenswrapper[4998]: E1203 16:22:11.064633 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.32:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Dec 03 16:22:11 crc kubenswrapper[4998]: E1203 16:22:11.064686 4998 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.32:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Dec 03 16:22:11 crc kubenswrapper[4998]: E1203 16:22:11.064842 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.32:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kwkfk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-77b567bfc7-g5znf_openstack(8d2d7661-f999-4c89-9e97-dce07062c850): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 16:22:11 crc kubenswrapper[4998]: E1203 16:22:11.065966 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-77b567bfc7-g5znf" podUID="8d2d7661-f999-4c89-9e97-dce07062c850" Dec 03 16:22:11 crc kubenswrapper[4998]: E1203 16:22:11.066039 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.32:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Dec 03 16:22:11 crc kubenswrapper[4998]: E1203 16:22:11.066063 4998 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.32:5001/podified-master-centos10/openstack-neutron-server:watcher_latest" Dec 03 16:22:11 crc kubenswrapper[4998]: E1203 16:22:11.066166 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:38.102.83.32:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fmfd9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-548dc7894c-6vwpc_openstack(3e93a28e-5430-4fac-8948-67ddf8db3a87): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 16:22:11 crc kubenswrapper[4998]: E1203 16:22:11.067260 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-548dc7894c-6vwpc" podUID="3e93a28e-5430-4fac-8948-67ddf8db3a87" Dec 03 16:22:11 crc kubenswrapper[4998]: W1203 16:22:11.080369 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0ef6669_5a8b_462d_9681_925859c86df1.slice/crio-7374ae0bc9a1d5462c5c581d373c9fc36e39a83f16d0e21c5e3d77f0bb83eb1d WatchSource:0}: Error finding container 7374ae0bc9a1d5462c5c581d373c9fc36e39a83f16d0e21c5e3d77f0bb83eb1d: Status 404 returned error can't find the container with id 7374ae0bc9a1d5462c5c581d373c9fc36e39a83f16d0e21c5e3d77f0bb83eb1d Dec 03 16:22:11 crc kubenswrapper[4998]: I1203 16:22:11.175614 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e0ef6669-5a8b-462d-9681-925859c86df1","Type":"ContainerStarted","Data":"7374ae0bc9a1d5462c5c581d373c9fc36e39a83f16d0e21c5e3d77f0bb83eb1d"} Dec 03 16:22:11 crc kubenswrapper[4998]: I1203 16:22:11.180545 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"aeb4a971-3c03-457c-b0df-1503701aac5b","Type":"ContainerStarted","Data":"4b6f937511621fe03bbf3a85fbbb4a9d458cac75832cdb13200fa423199deea7"} Dec 03 16:22:11 crc kubenswrapper[4998]: I1203 16:22:11.182803 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7786d8fd7-fmq4c" event={"ID":"27c5a979-99f8-46a9-9a96-312c8c61186f","Type":"ContainerStarted","Data":"3fc343c2c1818f5d350875851d70b58b6ff0858d74aed9192c757679e6d36d42"} Dec 03 16:22:11 crc kubenswrapper[4998]: I1203 16:22:11.844899 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-548dc7894c-6vwpc" Dec 03 16:22:11 crc kubenswrapper[4998]: I1203 16:22:11.851547 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77b567bfc7-g5znf" Dec 03 16:22:11 crc kubenswrapper[4998]: I1203 16:22:11.928582 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwkfk\" (UniqueName: \"kubernetes.io/projected/8d2d7661-f999-4c89-9e97-dce07062c850-kube-api-access-kwkfk\") pod \"8d2d7661-f999-4c89-9e97-dce07062c850\" (UID: \"8d2d7661-f999-4c89-9e97-dce07062c850\") " Dec 03 16:22:11 crc kubenswrapper[4998]: I1203 16:22:11.928669 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmfd9\" (UniqueName: \"kubernetes.io/projected/3e93a28e-5430-4fac-8948-67ddf8db3a87-kube-api-access-fmfd9\") pod \"3e93a28e-5430-4fac-8948-67ddf8db3a87\" (UID: \"3e93a28e-5430-4fac-8948-67ddf8db3a87\") " Dec 03 16:22:11 crc kubenswrapper[4998]: I1203 16:22:11.928729 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d2d7661-f999-4c89-9e97-dce07062c850-dns-svc\") pod \"8d2d7661-f999-4c89-9e97-dce07062c850\" (UID: \"8d2d7661-f999-4c89-9e97-dce07062c850\") " Dec 03 16:22:11 crc kubenswrapper[4998]: I1203 16:22:11.928782 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e93a28e-5430-4fac-8948-67ddf8db3a87-config\") pod \"3e93a28e-5430-4fac-8948-67ddf8db3a87\" (UID: \"3e93a28e-5430-4fac-8948-67ddf8db3a87\") " Dec 03 16:22:11 crc kubenswrapper[4998]: I1203 16:22:11.928825 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d2d7661-f999-4c89-9e97-dce07062c850-config\") pod \"8d2d7661-f999-4c89-9e97-dce07062c850\" (UID: \"8d2d7661-f999-4c89-9e97-dce07062c850\") " Dec 03 16:22:11 crc kubenswrapper[4998]: I1203 16:22:11.929871 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d2d7661-f999-4c89-9e97-dce07062c850-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8d2d7661-f999-4c89-9e97-dce07062c850" (UID: "8d2d7661-f999-4c89-9e97-dce07062c850"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:22:11 crc kubenswrapper[4998]: I1203 16:22:11.930027 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e93a28e-5430-4fac-8948-67ddf8db3a87-config" (OuterVolumeSpecName: "config") pod "3e93a28e-5430-4fac-8948-67ddf8db3a87" (UID: "3e93a28e-5430-4fac-8948-67ddf8db3a87"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:22:11 crc kubenswrapper[4998]: I1203 16:22:11.930076 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d2d7661-f999-4c89-9e97-dce07062c850-config" (OuterVolumeSpecName: "config") pod "8d2d7661-f999-4c89-9e97-dce07062c850" (UID: "8d2d7661-f999-4c89-9e97-dce07062c850"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:22:11 crc kubenswrapper[4998]: I1203 16:22:11.934610 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d2d7661-f999-4c89-9e97-dce07062c850-kube-api-access-kwkfk" (OuterVolumeSpecName: "kube-api-access-kwkfk") pod "8d2d7661-f999-4c89-9e97-dce07062c850" (UID: "8d2d7661-f999-4c89-9e97-dce07062c850"). InnerVolumeSpecName "kube-api-access-kwkfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:22:11 crc kubenswrapper[4998]: I1203 16:22:11.946308 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e93a28e-5430-4fac-8948-67ddf8db3a87-kube-api-access-fmfd9" (OuterVolumeSpecName: "kube-api-access-fmfd9") pod "3e93a28e-5430-4fac-8948-67ddf8db3a87" (UID: "3e93a28e-5430-4fac-8948-67ddf8db3a87"). InnerVolumeSpecName "kube-api-access-fmfd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.031630 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwkfk\" (UniqueName: \"kubernetes.io/projected/8d2d7661-f999-4c89-9e97-dce07062c850-kube-api-access-kwkfk\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.031674 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmfd9\" (UniqueName: \"kubernetes.io/projected/3e93a28e-5430-4fac-8948-67ddf8db3a87-kube-api-access-fmfd9\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.031690 4998 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8d2d7661-f999-4c89-9e97-dce07062c850-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.031703 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e93a28e-5430-4fac-8948-67ddf8db3a87-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.031717 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8d2d7661-f999-4c89-9e97-dce07062c850-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.141458 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-594b65fc49-cgfb2"] Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.156989 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.165227 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-7sxbw"] Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.175237 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 16:22:12 crc kubenswrapper[4998]: W1203 16:22:12.179279 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b2c3df0_703c_438e_9887_cf29844b9eb7.slice/crio-f90b2822a63bb8a04cae2d24f527b4e3e9ba07197a276e2aecf47b60044c1cbf WatchSource:0}: Error finding container f90b2822a63bb8a04cae2d24f527b4e3e9ba07197a276e2aecf47b60044c1cbf: Status 404 returned error can't find the container with id f90b2822a63bb8a04cae2d24f527b4e3e9ba07197a276e2aecf47b60044c1cbf Dec 03 16:22:12 crc kubenswrapper[4998]: W1203 16:22:12.180710 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ea3f8b4_86fc_4888_b9c0_36859ab25724.slice/crio-6e4fe1496bdf5e32bcb894a7a72bfc407a22150229b129938a632326c1c938af WatchSource:0}: Error finding container 6e4fe1496bdf5e32bcb894a7a72bfc407a22150229b129938a632326c1c938af: Status 404 returned error can't find the container with id 6e4fe1496bdf5e32bcb894a7a72bfc407a22150229b129938a632326c1c938af Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.183589 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.193542 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-notifications-server-0"] Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.196447 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8ea3f8b4-86fc-4888-b9c0-36859ab25724","Type":"ContainerStarted","Data":"6e4fe1496bdf5e32bcb894a7a72bfc407a22150229b129938a632326c1c938af"} Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.199882 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-548dc7894c-6vwpc" event={"ID":"3e93a28e-5430-4fac-8948-67ddf8db3a87","Type":"ContainerDied","Data":"4b2d63cfa91e30b46976fa9b56c9da82777e3d470ec0f093788aeabd6e7a0783"} Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.199975 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-548dc7894c-6vwpc" Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.200995 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.203709 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594b65fc49-cgfb2" event={"ID":"39e12872-2efe-4cec-82dd-a265f58da9ec","Type":"ContainerStarted","Data":"e7d096a06ca4e1c5aa1de8f9a6a65506e7a52af068a47625f35ad98492a46371"} Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.206722 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77b567bfc7-g5znf" event={"ID":"8d2d7661-f999-4c89-9e97-dce07062c850","Type":"ContainerDied","Data":"88caefdb3437a280bdcd0dd9e91b9522cb285d60db2623e2f92e459a2ad3fc44"} Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.206840 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77b567bfc7-g5znf" Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.208171 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.213628 4998 generic.go:334] "Generic (PLEG): container finished" podID="27c5a979-99f8-46a9-9a96-312c8c61186f" containerID="d51f2297cf0d8f82d6a221d7ec19a91a1af35082e627ac9c26314008339d3036" exitCode=0 Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.214406 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7786d8fd7-fmq4c" event={"ID":"27c5a979-99f8-46a9-9a96-312c8c61186f","Type":"ContainerDied","Data":"d51f2297cf0d8f82d6a221d7ec19a91a1af35082e627ac9c26314008339d3036"} Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.236842 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7sxbw" event={"ID":"95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f","Type":"ContainerStarted","Data":"f410a90b8baaf4fdbdf43612c89403b6c097093fc1b747c3ed85684eeeede726"} Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.239730 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92d2d9f9-b0c3-40f8-981a-41ec47f030e1","Type":"ContainerStarted","Data":"85e455c8ca11f61be037396af6a3f899fc6743b182bf018a7a79b9b2f7d475e2"} Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.386127 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76c44b4bf7-v5mhz"] Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.455284 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77b567bfc7-g5znf"] Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.467195 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77b567bfc7-g5znf"] Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.487473 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-548dc7894c-6vwpc"] Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.497872 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-548dc7894c-6vwpc"] Dec 03 16:22:12 crc kubenswrapper[4998]: I1203 16:22:12.497912 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 16:22:12 crc kubenswrapper[4998]: W1203 16:22:12.529335 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37aa6aee_15f8_4e23_a18c_216b85e605cf.slice/crio-74ee1e0defd71783e2fc95ecd164d0c20d686de3c3a5977fc00124d110061625 WatchSource:0}: Error finding container 74ee1e0defd71783e2fc95ecd164d0c20d686de3c3a5977fc00124d110061625: Status 404 returned error can't find the container with id 74ee1e0defd71783e2fc95ecd164d0c20d686de3c3a5977fc00124d110061625 Dec 03 16:22:12 crc kubenswrapper[4998]: E1203 16:22:12.656387 4998 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Dec 03 16:22:12 crc kubenswrapper[4998]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/27c5a979-99f8-46a9-9a96-312c8c61186f/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 03 16:22:12 crc kubenswrapper[4998]: > podSandboxID="3fc343c2c1818f5d350875851d70b58b6ff0858d74aed9192c757679e6d36d42" Dec 03 16:22:12 crc kubenswrapper[4998]: E1203 16:22:12.656836 4998 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 03 16:22:12 crc kubenswrapper[4998]: container &Container{Name:dnsmasq-dns,Image:38.102.83.32:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v5pdb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7786d8fd7-fmq4c_openstack(27c5a979-99f8-46a9-9a96-312c8c61186f): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/27c5a979-99f8-46a9-9a96-312c8c61186f/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 03 16:22:12 crc kubenswrapper[4998]: > logger="UnhandledError" Dec 03 16:22:12 crc kubenswrapper[4998]: E1203 16:22:12.657979 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/27c5a979-99f8-46a9-9a96-312c8c61186f/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-7786d8fd7-fmq4c" podUID="27c5a979-99f8-46a9-9a96-312c8c61186f" Dec 03 16:22:13 crc kubenswrapper[4998]: I1203 16:22:13.254395 4998 generic.go:334] "Generic (PLEG): container finished" podID="39e12872-2efe-4cec-82dd-a265f58da9ec" containerID="5254702d498c1d57539ad9a53df0e92c424d723090b3e059832ad451b99efb5f" exitCode=0 Dec 03 16:22:13 crc kubenswrapper[4998]: I1203 16:22:13.254485 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594b65fc49-cgfb2" event={"ID":"39e12872-2efe-4cec-82dd-a265f58da9ec","Type":"ContainerDied","Data":"5254702d498c1d57539ad9a53df0e92c424d723090b3e059832ad451b99efb5f"} Dec 03 16:22:13 crc kubenswrapper[4998]: I1203 16:22:13.256189 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"dbde2166-b545-4979-845a-8932845b9cd1","Type":"ContainerStarted","Data":"380ea97281db5192bbecf694161b3dcd2bf3d34bf20b95d3e8ad743af8636b19"} Dec 03 16:22:13 crc kubenswrapper[4998]: I1203 16:22:13.258639 4998 generic.go:334] "Generic (PLEG): container finished" podID="c0c40a4e-c357-48d4-a3e1-a1eb8ba85256" containerID="3a009f00ee321f31833c4020232cd1ed7dbcaf3e41053457f1f789e6db656ee3" exitCode=0 Dec 03 16:22:13 crc kubenswrapper[4998]: I1203 16:22:13.258693 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c44b4bf7-v5mhz" event={"ID":"c0c40a4e-c357-48d4-a3e1-a1eb8ba85256","Type":"ContainerDied","Data":"3a009f00ee321f31833c4020232cd1ed7dbcaf3e41053457f1f789e6db656ee3"} Dec 03 16:22:13 crc kubenswrapper[4998]: I1203 16:22:13.258712 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c44b4bf7-v5mhz" event={"ID":"c0c40a4e-c357-48d4-a3e1-a1eb8ba85256","Type":"ContainerStarted","Data":"2bc6ec238c6e293b899e1b2fc9714ed3218b9253feff2e05fa0ae55c17db4601"} Dec 03 16:22:13 crc kubenswrapper[4998]: I1203 16:22:13.261019 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f9abb1de-7121-4de9-833a-9f1ab889b8ff","Type":"ContainerStarted","Data":"dc5aac4b616dd2f5fbb9ca9a9c8ee062f5835f56ad689e0c9da7e0507b7a8a6f"} Dec 03 16:22:13 crc kubenswrapper[4998]: I1203 16:22:13.262298 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"75024598-618b-49e3-919c-3bf92e4251a2","Type":"ContainerStarted","Data":"d32184d40b7557b03a3e606e29dd6f22e969c7ece4bc51ad067fb40ea8266a79"} Dec 03 16:22:13 crc kubenswrapper[4998]: I1203 16:22:13.263480 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"37aa6aee-15f8-4e23-a18c-216b85e605cf","Type":"ContainerStarted","Data":"74ee1e0defd71783e2fc95ecd164d0c20d686de3c3a5977fc00124d110061625"} Dec 03 16:22:13 crc kubenswrapper[4998]: I1203 16:22:13.266003 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1b2c3df0-703c-438e-9887-cf29844b9eb7","Type":"ContainerStarted","Data":"f90b2822a63bb8a04cae2d24f527b4e3e9ba07197a276e2aecf47b60044c1cbf"} Dec 03 16:22:13 crc kubenswrapper[4998]: I1203 16:22:13.405142 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-5hxf2"] Dec 03 16:22:13 crc kubenswrapper[4998]: I1203 16:22:13.558237 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 16:22:13 crc kubenswrapper[4998]: I1203 16:22:13.689941 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e93a28e-5430-4fac-8948-67ddf8db3a87" path="/var/lib/kubelet/pods/3e93a28e-5430-4fac-8948-67ddf8db3a87/volumes" Dec 03 16:22:13 crc kubenswrapper[4998]: I1203 16:22:13.690368 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d2d7661-f999-4c89-9e97-dce07062c850" path="/var/lib/kubelet/pods/8d2d7661-f999-4c89-9e97-dce07062c850/volumes" Dec 03 16:22:17 crc kubenswrapper[4998]: W1203 16:22:17.194355 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod893cf12f_77a9_463f_8dd4_1b3f05b2c9bf.slice/crio-1734382337e48f0a2db1d73a2278f4627f71e7ff193eba8dc3d0e4765b7bc8b8 WatchSource:0}: Error finding container 1734382337e48f0a2db1d73a2278f4627f71e7ff193eba8dc3d0e4765b7bc8b8: Status 404 returned error can't find the container with id 1734382337e48f0a2db1d73a2278f4627f71e7ff193eba8dc3d0e4765b7bc8b8 Dec 03 16:22:17 crc kubenswrapper[4998]: W1203 16:22:17.196878 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod832cc7cb_044b_4bb9_84d9_74e3d5a8f4a5.slice/crio-8970789a3b2f650622d7fc5645f966514c3260d0a03fe027cc59370b11d2b10b WatchSource:0}: Error finding container 8970789a3b2f650622d7fc5645f966514c3260d0a03fe027cc59370b11d2b10b: Status 404 returned error can't find the container with id 8970789a3b2f650622d7fc5645f966514c3260d0a03fe027cc59370b11d2b10b Dec 03 16:22:17 crc kubenswrapper[4998]: I1203 16:22:17.261450 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76c44b4bf7-v5mhz" Dec 03 16:22:17 crc kubenswrapper[4998]: I1203 16:22:17.300674 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5hxf2" event={"ID":"832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5","Type":"ContainerStarted","Data":"8970789a3b2f650622d7fc5645f966514c3260d0a03fe027cc59370b11d2b10b"} Dec 03 16:22:17 crc kubenswrapper[4998]: I1203 16:22:17.302148 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf","Type":"ContainerStarted","Data":"1734382337e48f0a2db1d73a2278f4627f71e7ff193eba8dc3d0e4765b7bc8b8"} Dec 03 16:22:17 crc kubenswrapper[4998]: I1203 16:22:17.304136 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c44b4bf7-v5mhz" event={"ID":"c0c40a4e-c357-48d4-a3e1-a1eb8ba85256","Type":"ContainerDied","Data":"2bc6ec238c6e293b899e1b2fc9714ed3218b9253feff2e05fa0ae55c17db4601"} Dec 03 16:22:17 crc kubenswrapper[4998]: I1203 16:22:17.304202 4998 scope.go:117] "RemoveContainer" containerID="3a009f00ee321f31833c4020232cd1ed7dbcaf3e41053457f1f789e6db656ee3" Dec 03 16:22:17 crc kubenswrapper[4998]: I1203 16:22:17.304254 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76c44b4bf7-v5mhz" Dec 03 16:22:17 crc kubenswrapper[4998]: I1203 16:22:17.328074 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c0c40a4e-c357-48d4-a3e1-a1eb8ba85256-dns-svc\") pod \"c0c40a4e-c357-48d4-a3e1-a1eb8ba85256\" (UID: \"c0c40a4e-c357-48d4-a3e1-a1eb8ba85256\") " Dec 03 16:22:17 crc kubenswrapper[4998]: I1203 16:22:17.328236 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzvgk\" (UniqueName: \"kubernetes.io/projected/c0c40a4e-c357-48d4-a3e1-a1eb8ba85256-kube-api-access-kzvgk\") pod \"c0c40a4e-c357-48d4-a3e1-a1eb8ba85256\" (UID: \"c0c40a4e-c357-48d4-a3e1-a1eb8ba85256\") " Dec 03 16:22:17 crc kubenswrapper[4998]: I1203 16:22:17.328262 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0c40a4e-c357-48d4-a3e1-a1eb8ba85256-config\") pod \"c0c40a4e-c357-48d4-a3e1-a1eb8ba85256\" (UID: \"c0c40a4e-c357-48d4-a3e1-a1eb8ba85256\") " Dec 03 16:22:17 crc kubenswrapper[4998]: I1203 16:22:17.334850 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0c40a4e-c357-48d4-a3e1-a1eb8ba85256-kube-api-access-kzvgk" (OuterVolumeSpecName: "kube-api-access-kzvgk") pod "c0c40a4e-c357-48d4-a3e1-a1eb8ba85256" (UID: "c0c40a4e-c357-48d4-a3e1-a1eb8ba85256"). InnerVolumeSpecName "kube-api-access-kzvgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:22:17 crc kubenswrapper[4998]: I1203 16:22:17.349703 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0c40a4e-c357-48d4-a3e1-a1eb8ba85256-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c0c40a4e-c357-48d4-a3e1-a1eb8ba85256" (UID: "c0c40a4e-c357-48d4-a3e1-a1eb8ba85256"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:22:17 crc kubenswrapper[4998]: I1203 16:22:17.350587 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0c40a4e-c357-48d4-a3e1-a1eb8ba85256-config" (OuterVolumeSpecName: "config") pod "c0c40a4e-c357-48d4-a3e1-a1eb8ba85256" (UID: "c0c40a4e-c357-48d4-a3e1-a1eb8ba85256"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:22:17 crc kubenswrapper[4998]: I1203 16:22:17.429599 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzvgk\" (UniqueName: \"kubernetes.io/projected/c0c40a4e-c357-48d4-a3e1-a1eb8ba85256-kube-api-access-kzvgk\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:17 crc kubenswrapper[4998]: I1203 16:22:17.429631 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0c40a4e-c357-48d4-a3e1-a1eb8ba85256-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:17 crc kubenswrapper[4998]: I1203 16:22:17.429642 4998 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c0c40a4e-c357-48d4-a3e1-a1eb8ba85256-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:17 crc kubenswrapper[4998]: I1203 16:22:17.656057 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76c44b4bf7-v5mhz"] Dec 03 16:22:17 crc kubenswrapper[4998]: I1203 16:22:17.662187 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76c44b4bf7-v5mhz"] Dec 03 16:22:17 crc kubenswrapper[4998]: I1203 16:22:17.688579 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0c40a4e-c357-48d4-a3e1-a1eb8ba85256" path="/var/lib/kubelet/pods/c0c40a4e-c357-48d4-a3e1-a1eb8ba85256/volumes" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.220123 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-w7sx4"] Dec 03 16:22:23 crc kubenswrapper[4998]: E1203 16:22:23.225038 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0c40a4e-c357-48d4-a3e1-a1eb8ba85256" containerName="init" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.225077 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0c40a4e-c357-48d4-a3e1-a1eb8ba85256" containerName="init" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.225270 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0c40a4e-c357-48d4-a3e1-a1eb8ba85256" containerName="init" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.225905 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-w7sx4" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.231259 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.243662 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-w7sx4"] Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.333356 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebaa6f1f-170e-41a4-ad4d-332108a0233b-combined-ca-bundle\") pod \"ovn-controller-metrics-w7sx4\" (UID: \"ebaa6f1f-170e-41a4-ad4d-332108a0233b\") " pod="openstack/ovn-controller-metrics-w7sx4" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.333608 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n6jk\" (UniqueName: \"kubernetes.io/projected/ebaa6f1f-170e-41a4-ad4d-332108a0233b-kube-api-access-6n6jk\") pod \"ovn-controller-metrics-w7sx4\" (UID: \"ebaa6f1f-170e-41a4-ad4d-332108a0233b\") " pod="openstack/ovn-controller-metrics-w7sx4" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.333631 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebaa6f1f-170e-41a4-ad4d-332108a0233b-config\") pod \"ovn-controller-metrics-w7sx4\" (UID: \"ebaa6f1f-170e-41a4-ad4d-332108a0233b\") " pod="openstack/ovn-controller-metrics-w7sx4" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.333715 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ebaa6f1f-170e-41a4-ad4d-332108a0233b-ovs-rundir\") pod \"ovn-controller-metrics-w7sx4\" (UID: \"ebaa6f1f-170e-41a4-ad4d-332108a0233b\") " pod="openstack/ovn-controller-metrics-w7sx4" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.333735 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ebaa6f1f-170e-41a4-ad4d-332108a0233b-ovn-rundir\") pod \"ovn-controller-metrics-w7sx4\" (UID: \"ebaa6f1f-170e-41a4-ad4d-332108a0233b\") " pod="openstack/ovn-controller-metrics-w7sx4" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.333779 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebaa6f1f-170e-41a4-ad4d-332108a0233b-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-w7sx4\" (UID: \"ebaa6f1f-170e-41a4-ad4d-332108a0233b\") " pod="openstack/ovn-controller-metrics-w7sx4" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.366334 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8ea3f8b4-86fc-4888-b9c0-36859ab25724","Type":"ContainerStarted","Data":"d65ca50f1c810cea0895251df4ced37c187aab806e40c39e892ccc774306b756"} Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.378652 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1b2c3df0-703c-438e-9887-cf29844b9eb7","Type":"ContainerStarted","Data":"6e37399ddc89c2d43698c7ae97f2b25c082586d474cceb3b7c586516d81af9fd"} Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.383092 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7786d8fd7-fmq4c"] Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.385390 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594b65fc49-cgfb2" event={"ID":"39e12872-2efe-4cec-82dd-a265f58da9ec","Type":"ContainerStarted","Data":"1cdb1825530aab9db06bbdc28a5a98ec78ab0055779a12059dbf2c4a0e0e57d2"} Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.386245 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-594b65fc49-cgfb2" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.414778 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7786d8fd7-fmq4c" event={"ID":"27c5a979-99f8-46a9-9a96-312c8c61186f","Type":"ContainerStarted","Data":"b9ec91c1bee81c5e74567c92b09c45e79715222f9b173a4083c26528bfd1ec04"} Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.414967 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7786d8fd7-fmq4c" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.414971 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7786d8fd7-fmq4c" podUID="27c5a979-99f8-46a9-9a96-312c8c61186f" containerName="dnsmasq-dns" containerID="cri-o://b9ec91c1bee81c5e74567c92b09c45e79715222f9b173a4083c26528bfd1ec04" gracePeriod=10 Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.424529 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b6fcbcc87-jjnql"] Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.427651 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b6fcbcc87-jjnql" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.432163 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.434738 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ebaa6f1f-170e-41a4-ad4d-332108a0233b-ovs-rundir\") pod \"ovn-controller-metrics-w7sx4\" (UID: \"ebaa6f1f-170e-41a4-ad4d-332108a0233b\") " pod="openstack/ovn-controller-metrics-w7sx4" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.434805 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ebaa6f1f-170e-41a4-ad4d-332108a0233b-ovn-rundir\") pod \"ovn-controller-metrics-w7sx4\" (UID: \"ebaa6f1f-170e-41a4-ad4d-332108a0233b\") " pod="openstack/ovn-controller-metrics-w7sx4" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.434838 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebaa6f1f-170e-41a4-ad4d-332108a0233b-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-w7sx4\" (UID: \"ebaa6f1f-170e-41a4-ad4d-332108a0233b\") " pod="openstack/ovn-controller-metrics-w7sx4" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.434887 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebaa6f1f-170e-41a4-ad4d-332108a0233b-combined-ca-bundle\") pod \"ovn-controller-metrics-w7sx4\" (UID: \"ebaa6f1f-170e-41a4-ad4d-332108a0233b\") " pod="openstack/ovn-controller-metrics-w7sx4" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.434902 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n6jk\" (UniqueName: \"kubernetes.io/projected/ebaa6f1f-170e-41a4-ad4d-332108a0233b-kube-api-access-6n6jk\") pod \"ovn-controller-metrics-w7sx4\" (UID: \"ebaa6f1f-170e-41a4-ad4d-332108a0233b\") " pod="openstack/ovn-controller-metrics-w7sx4" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.434920 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebaa6f1f-170e-41a4-ad4d-332108a0233b-config\") pod \"ovn-controller-metrics-w7sx4\" (UID: \"ebaa6f1f-170e-41a4-ad4d-332108a0233b\") " pod="openstack/ovn-controller-metrics-w7sx4" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.435726 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ebaa6f1f-170e-41a4-ad4d-332108a0233b-ovn-rundir\") pod \"ovn-controller-metrics-w7sx4\" (UID: \"ebaa6f1f-170e-41a4-ad4d-332108a0233b\") " pod="openstack/ovn-controller-metrics-w7sx4" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.435816 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ebaa6f1f-170e-41a4-ad4d-332108a0233b-ovs-rundir\") pod \"ovn-controller-metrics-w7sx4\" (UID: \"ebaa6f1f-170e-41a4-ad4d-332108a0233b\") " pod="openstack/ovn-controller-metrics-w7sx4" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.437194 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebaa6f1f-170e-41a4-ad4d-332108a0233b-config\") pod \"ovn-controller-metrics-w7sx4\" (UID: \"ebaa6f1f-170e-41a4-ad4d-332108a0233b\") " pod="openstack/ovn-controller-metrics-w7sx4" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.445904 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebaa6f1f-170e-41a4-ad4d-332108a0233b-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-w7sx4\" (UID: \"ebaa6f1f-170e-41a4-ad4d-332108a0233b\") " pod="openstack/ovn-controller-metrics-w7sx4" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.459135 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebaa6f1f-170e-41a4-ad4d-332108a0233b-combined-ca-bundle\") pod \"ovn-controller-metrics-w7sx4\" (UID: \"ebaa6f1f-170e-41a4-ad4d-332108a0233b\") " pod="openstack/ovn-controller-metrics-w7sx4" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.459305 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b6fcbcc87-jjnql"] Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.474321 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n6jk\" (UniqueName: \"kubernetes.io/projected/ebaa6f1f-170e-41a4-ad4d-332108a0233b-kube-api-access-6n6jk\") pod \"ovn-controller-metrics-w7sx4\" (UID: \"ebaa6f1f-170e-41a4-ad4d-332108a0233b\") " pod="openstack/ovn-controller-metrics-w7sx4" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.475638 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-594b65fc49-cgfb2" podStartSLOduration=34.475621197 podStartE2EDuration="34.475621197s" podCreationTimestamp="2025-12-03 16:21:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:22:23.436496031 +0000 UTC m=+1122.048196254" watchObservedRunningTime="2025-12-03 16:22:23.475621197 +0000 UTC m=+1122.087321420" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.514090 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7786d8fd7-fmq4c" podStartSLOduration=34.585105723 podStartE2EDuration="35.514075056s" podCreationTimestamp="2025-12-03 16:21:48 +0000 UTC" firstStartedPulling="2025-12-03 16:22:10.245216182 +0000 UTC m=+1108.856916405" lastFinishedPulling="2025-12-03 16:22:11.174185515 +0000 UTC m=+1109.785885738" observedRunningTime="2025-12-03 16:22:23.510414844 +0000 UTC m=+1122.122115067" watchObservedRunningTime="2025-12-03 16:22:23.514075056 +0000 UTC m=+1122.125775279" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.559073 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-w7sx4" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.577325 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-594b65fc49-cgfb2"] Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.611813 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56c4b567fc-78kqx"] Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.613237 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.618046 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.623491 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56c4b567fc-78kqx"] Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.642283 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c6b2505-697f-4935-af28-88a4ae94524f-dns-svc\") pod \"dnsmasq-dns-5b6fcbcc87-jjnql\" (UID: \"4c6b2505-697f-4935-af28-88a4ae94524f\") " pod="openstack/dnsmasq-dns-5b6fcbcc87-jjnql" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.642365 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c6b2505-697f-4935-af28-88a4ae94524f-config\") pod \"dnsmasq-dns-5b6fcbcc87-jjnql\" (UID: \"4c6b2505-697f-4935-af28-88a4ae94524f\") " pod="openstack/dnsmasq-dns-5b6fcbcc87-jjnql" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.642388 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzwzl\" (UniqueName: \"kubernetes.io/projected/4c6b2505-697f-4935-af28-88a4ae94524f-kube-api-access-dzwzl\") pod \"dnsmasq-dns-5b6fcbcc87-jjnql\" (UID: \"4c6b2505-697f-4935-af28-88a4ae94524f\") " pod="openstack/dnsmasq-dns-5b6fcbcc87-jjnql" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.642433 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c6b2505-697f-4935-af28-88a4ae94524f-ovsdbserver-nb\") pod \"dnsmasq-dns-5b6fcbcc87-jjnql\" (UID: \"4c6b2505-697f-4935-af28-88a4ae94524f\") " pod="openstack/dnsmasq-dns-5b6fcbcc87-jjnql" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.744124 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-ovsdbserver-sb\") pod \"dnsmasq-dns-56c4b567fc-78kqx\" (UID: \"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580\") " pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.744171 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgf84\" (UniqueName: \"kubernetes.io/projected/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-kube-api-access-dgf84\") pod \"dnsmasq-dns-56c4b567fc-78kqx\" (UID: \"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580\") " pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.744207 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c6b2505-697f-4935-af28-88a4ae94524f-config\") pod \"dnsmasq-dns-5b6fcbcc87-jjnql\" (UID: \"4c6b2505-697f-4935-af28-88a4ae94524f\") " pod="openstack/dnsmasq-dns-5b6fcbcc87-jjnql" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.744255 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzwzl\" (UniqueName: \"kubernetes.io/projected/4c6b2505-697f-4935-af28-88a4ae94524f-kube-api-access-dzwzl\") pod \"dnsmasq-dns-5b6fcbcc87-jjnql\" (UID: \"4c6b2505-697f-4935-af28-88a4ae94524f\") " pod="openstack/dnsmasq-dns-5b6fcbcc87-jjnql" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.744293 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-ovsdbserver-nb\") pod \"dnsmasq-dns-56c4b567fc-78kqx\" (UID: \"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580\") " pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.744331 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-config\") pod \"dnsmasq-dns-56c4b567fc-78kqx\" (UID: \"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580\") " pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.744356 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c6b2505-697f-4935-af28-88a4ae94524f-ovsdbserver-nb\") pod \"dnsmasq-dns-5b6fcbcc87-jjnql\" (UID: \"4c6b2505-697f-4935-af28-88a4ae94524f\") " pod="openstack/dnsmasq-dns-5b6fcbcc87-jjnql" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.744381 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-dns-svc\") pod \"dnsmasq-dns-56c4b567fc-78kqx\" (UID: \"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580\") " pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.744565 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c6b2505-697f-4935-af28-88a4ae94524f-dns-svc\") pod \"dnsmasq-dns-5b6fcbcc87-jjnql\" (UID: \"4c6b2505-697f-4935-af28-88a4ae94524f\") " pod="openstack/dnsmasq-dns-5b6fcbcc87-jjnql" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.745118 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c6b2505-697f-4935-af28-88a4ae94524f-ovsdbserver-nb\") pod \"dnsmasq-dns-5b6fcbcc87-jjnql\" (UID: \"4c6b2505-697f-4935-af28-88a4ae94524f\") " pod="openstack/dnsmasq-dns-5b6fcbcc87-jjnql" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.745263 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c6b2505-697f-4935-af28-88a4ae94524f-config\") pod \"dnsmasq-dns-5b6fcbcc87-jjnql\" (UID: \"4c6b2505-697f-4935-af28-88a4ae94524f\") " pod="openstack/dnsmasq-dns-5b6fcbcc87-jjnql" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.745511 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c6b2505-697f-4935-af28-88a4ae94524f-dns-svc\") pod \"dnsmasq-dns-5b6fcbcc87-jjnql\" (UID: \"4c6b2505-697f-4935-af28-88a4ae94524f\") " pod="openstack/dnsmasq-dns-5b6fcbcc87-jjnql" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.804649 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzwzl\" (UniqueName: \"kubernetes.io/projected/4c6b2505-697f-4935-af28-88a4ae94524f-kube-api-access-dzwzl\") pod \"dnsmasq-dns-5b6fcbcc87-jjnql\" (UID: \"4c6b2505-697f-4935-af28-88a4ae94524f\") " pod="openstack/dnsmasq-dns-5b6fcbcc87-jjnql" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.846411 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-ovsdbserver-sb\") pod \"dnsmasq-dns-56c4b567fc-78kqx\" (UID: \"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580\") " pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.846461 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgf84\" (UniqueName: \"kubernetes.io/projected/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-kube-api-access-dgf84\") pod \"dnsmasq-dns-56c4b567fc-78kqx\" (UID: \"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580\") " pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.846504 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-ovsdbserver-nb\") pod \"dnsmasq-dns-56c4b567fc-78kqx\" (UID: \"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580\") " pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.846545 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-config\") pod \"dnsmasq-dns-56c4b567fc-78kqx\" (UID: \"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580\") " pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.846574 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-dns-svc\") pod \"dnsmasq-dns-56c4b567fc-78kqx\" (UID: \"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580\") " pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.847405 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-ovsdbserver-sb\") pod \"dnsmasq-dns-56c4b567fc-78kqx\" (UID: \"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580\") " pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.847517 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-dns-svc\") pod \"dnsmasq-dns-56c4b567fc-78kqx\" (UID: \"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580\") " pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.847543 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-config\") pod \"dnsmasq-dns-56c4b567fc-78kqx\" (UID: \"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580\") " pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.848153 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-ovsdbserver-nb\") pod \"dnsmasq-dns-56c4b567fc-78kqx\" (UID: \"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580\") " pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" Dec 03 16:22:23 crc kubenswrapper[4998]: I1203 16:22:23.864510 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgf84\" (UniqueName: \"kubernetes.io/projected/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-kube-api-access-dgf84\") pod \"dnsmasq-dns-56c4b567fc-78kqx\" (UID: \"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580\") " pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" Dec 03 16:22:24 crc kubenswrapper[4998]: I1203 16:22:24.302745 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b6fcbcc87-jjnql" Dec 03 16:22:24 crc kubenswrapper[4998]: I1203 16:22:24.315467 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" Dec 03 16:22:24 crc kubenswrapper[4998]: I1203 16:22:24.426527 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5hxf2" event={"ID":"832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5","Type":"ContainerStarted","Data":"c6c2d407845abea567991ba62ec1ef16217061981777d2c8d2d5f28527b8dfad"} Dec 03 16:22:24 crc kubenswrapper[4998]: I1203 16:22:24.427993 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7sxbw" event={"ID":"95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f","Type":"ContainerStarted","Data":"711e2a4e9f7b67cef73da005d7ba400adec929b0f0a0b65e456e4bcb702f0e56"} Dec 03 16:22:24 crc kubenswrapper[4998]: I1203 16:22:24.430057 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"75024598-618b-49e3-919c-3bf92e4251a2","Type":"ContainerStarted","Data":"22d58e4b6f5c02909d29ed0a58a18e1962d58e0f46897dac93b124d8cde8ad75"} Dec 03 16:22:24 crc kubenswrapper[4998]: I1203 16:22:24.431782 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"37aa6aee-15f8-4e23-a18c-216b85e605cf","Type":"ContainerStarted","Data":"3b0a1a53b0b961c38ce3cc9aee44b33203d288ed61b7cf666da70eac05168e75"} Dec 03 16:22:25 crc kubenswrapper[4998]: I1203 16:22:25.440881 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"dbde2166-b545-4979-845a-8932845b9cd1","Type":"ContainerStarted","Data":"5ab3b6bc5babd3b03187dd12c421440a4984c2949d620b458c03048793037989"} Dec 03 16:22:25 crc kubenswrapper[4998]: I1203 16:22:25.443733 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"aeb4a971-3c03-457c-b0df-1503701aac5b","Type":"ContainerStarted","Data":"d9927260898b6eed6b91ad5a3c179db33be6ad3449ab3cb52a37f8296becfaed"} Dec 03 16:22:25 crc kubenswrapper[4998]: I1203 16:22:25.447250 4998 generic.go:334] "Generic (PLEG): container finished" podID="27c5a979-99f8-46a9-9a96-312c8c61186f" containerID="b9ec91c1bee81c5e74567c92b09c45e79715222f9b173a4083c26528bfd1ec04" exitCode=0 Dec 03 16:22:25 crc kubenswrapper[4998]: I1203 16:22:25.447343 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7786d8fd7-fmq4c" event={"ID":"27c5a979-99f8-46a9-9a96-312c8c61186f","Type":"ContainerDied","Data":"b9ec91c1bee81c5e74567c92b09c45e79715222f9b173a4083c26528bfd1ec04"} Dec 03 16:22:25 crc kubenswrapper[4998]: I1203 16:22:25.450210 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e0ef6669-5a8b-462d-9681-925859c86df1","Type":"ContainerStarted","Data":"a3f53d09f4c25f953ec5e8b2e0a067502b460ce0068895898828046c063bc5c9"} Dec 03 16:22:25 crc kubenswrapper[4998]: I1203 16:22:25.450309 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 03 16:22:25 crc kubenswrapper[4998]: I1203 16:22:25.450333 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-594b65fc49-cgfb2" podUID="39e12872-2efe-4cec-82dd-a265f58da9ec" containerName="dnsmasq-dns" containerID="cri-o://1cdb1825530aab9db06bbdc28a5a98ec78ab0055779a12059dbf2c4a0e0e57d2" gracePeriod=10 Dec 03 16:22:25 crc kubenswrapper[4998]: I1203 16:22:25.450397 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-7sxbw" Dec 03 16:22:25 crc kubenswrapper[4998]: I1203 16:22:25.592266 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=21.283389939 podStartE2EDuration="31.592250401s" podCreationTimestamp="2025-12-03 16:21:54 +0000 UTC" firstStartedPulling="2025-12-03 16:22:12.210600316 +0000 UTC m=+1110.822300549" lastFinishedPulling="2025-12-03 16:22:22.519460748 +0000 UTC m=+1121.131161011" observedRunningTime="2025-12-03 16:22:25.586432945 +0000 UTC m=+1124.198133178" watchObservedRunningTime="2025-12-03 16:22:25.592250401 +0000 UTC m=+1124.203950614" Dec 03 16:22:25 crc kubenswrapper[4998]: I1203 16:22:25.613634 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-7sxbw" podStartSLOduration=17.003950071 podStartE2EDuration="26.613610689s" podCreationTimestamp="2025-12-03 16:21:59 +0000 UTC" firstStartedPulling="2025-12-03 16:22:12.165858638 +0000 UTC m=+1110.777558861" lastFinishedPulling="2025-12-03 16:22:21.775519216 +0000 UTC m=+1120.387219479" observedRunningTime="2025-12-03 16:22:25.603732291 +0000 UTC m=+1124.215432514" watchObservedRunningTime="2025-12-03 16:22:25.613610689 +0000 UTC m=+1124.225310922" Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.342349 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7786d8fd7-fmq4c" Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.353180 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-594b65fc49-cgfb2" Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.463747 4998 generic.go:334] "Generic (PLEG): container finished" podID="832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5" containerID="c6c2d407845abea567991ba62ec1ef16217061981777d2c8d2d5f28527b8dfad" exitCode=0 Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.464167 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5hxf2" event={"ID":"832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5","Type":"ContainerDied","Data":"c6c2d407845abea567991ba62ec1ef16217061981777d2c8d2d5f28527b8dfad"} Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.471219 4998 generic.go:334] "Generic (PLEG): container finished" podID="39e12872-2efe-4cec-82dd-a265f58da9ec" containerID="1cdb1825530aab9db06bbdc28a5a98ec78ab0055779a12059dbf2c4a0e0e57d2" exitCode=0 Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.471303 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594b65fc49-cgfb2" event={"ID":"39e12872-2efe-4cec-82dd-a265f58da9ec","Type":"ContainerDied","Data":"1cdb1825530aab9db06bbdc28a5a98ec78ab0055779a12059dbf2c4a0e0e57d2"} Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.471308 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-594b65fc49-cgfb2" Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.471332 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594b65fc49-cgfb2" event={"ID":"39e12872-2efe-4cec-82dd-a265f58da9ec","Type":"ContainerDied","Data":"e7d096a06ca4e1c5aa1de8f9a6a65506e7a52af068a47625f35ad98492a46371"} Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.471356 4998 scope.go:117] "RemoveContainer" containerID="1cdb1825530aab9db06bbdc28a5a98ec78ab0055779a12059dbf2c4a0e0e57d2" Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.476320 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7786d8fd7-fmq4c" event={"ID":"27c5a979-99f8-46a9-9a96-312c8c61186f","Type":"ContainerDied","Data":"3fc343c2c1818f5d350875851d70b58b6ff0858d74aed9192c757679e6d36d42"} Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.476392 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7786d8fd7-fmq4c" Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.480561 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92d2d9f9-b0c3-40f8-981a-41ec47f030e1","Type":"ContainerStarted","Data":"174a248e64d7d0441f1074334990836bdc3697b8f54e27015258af83a045fcd0"} Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.492165 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9j62\" (UniqueName: \"kubernetes.io/projected/39e12872-2efe-4cec-82dd-a265f58da9ec-kube-api-access-s9j62\") pod \"39e12872-2efe-4cec-82dd-a265f58da9ec\" (UID: \"39e12872-2efe-4cec-82dd-a265f58da9ec\") " Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.492245 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5pdb\" (UniqueName: \"kubernetes.io/projected/27c5a979-99f8-46a9-9a96-312c8c61186f-kube-api-access-v5pdb\") pod \"27c5a979-99f8-46a9-9a96-312c8c61186f\" (UID: \"27c5a979-99f8-46a9-9a96-312c8c61186f\") " Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.492291 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27c5a979-99f8-46a9-9a96-312c8c61186f-dns-svc\") pod \"27c5a979-99f8-46a9-9a96-312c8c61186f\" (UID: \"27c5a979-99f8-46a9-9a96-312c8c61186f\") " Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.492341 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39e12872-2efe-4cec-82dd-a265f58da9ec-config\") pod \"39e12872-2efe-4cec-82dd-a265f58da9ec\" (UID: \"39e12872-2efe-4cec-82dd-a265f58da9ec\") " Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.492411 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39e12872-2efe-4cec-82dd-a265f58da9ec-dns-svc\") pod \"39e12872-2efe-4cec-82dd-a265f58da9ec\" (UID: \"39e12872-2efe-4cec-82dd-a265f58da9ec\") " Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.492467 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27c5a979-99f8-46a9-9a96-312c8c61186f-config\") pod \"27c5a979-99f8-46a9-9a96-312c8c61186f\" (UID: \"27c5a979-99f8-46a9-9a96-312c8c61186f\") " Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.501382 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39e12872-2efe-4cec-82dd-a265f58da9ec-kube-api-access-s9j62" (OuterVolumeSpecName: "kube-api-access-s9j62") pod "39e12872-2efe-4cec-82dd-a265f58da9ec" (UID: "39e12872-2efe-4cec-82dd-a265f58da9ec"). InnerVolumeSpecName "kube-api-access-s9j62". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.502931 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27c5a979-99f8-46a9-9a96-312c8c61186f-kube-api-access-v5pdb" (OuterVolumeSpecName: "kube-api-access-v5pdb") pod "27c5a979-99f8-46a9-9a96-312c8c61186f" (UID: "27c5a979-99f8-46a9-9a96-312c8c61186f"). InnerVolumeSpecName "kube-api-access-v5pdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.589299 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39e12872-2efe-4cec-82dd-a265f58da9ec-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "39e12872-2efe-4cec-82dd-a265f58da9ec" (UID: "39e12872-2efe-4cec-82dd-a265f58da9ec"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.590028 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27c5a979-99f8-46a9-9a96-312c8c61186f-config" (OuterVolumeSpecName: "config") pod "27c5a979-99f8-46a9-9a96-312c8c61186f" (UID: "27c5a979-99f8-46a9-9a96-312c8c61186f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.593265 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27c5a979-99f8-46a9-9a96-312c8c61186f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "27c5a979-99f8-46a9-9a96-312c8c61186f" (UID: "27c5a979-99f8-46a9-9a96-312c8c61186f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.593315 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39e12872-2efe-4cec-82dd-a265f58da9ec-config" (OuterVolumeSpecName: "config") pod "39e12872-2efe-4cec-82dd-a265f58da9ec" (UID: "39e12872-2efe-4cec-82dd-a265f58da9ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.595194 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39e12872-2efe-4cec-82dd-a265f58da9ec-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.595214 4998 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39e12872-2efe-4cec-82dd-a265f58da9ec-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.595226 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27c5a979-99f8-46a9-9a96-312c8c61186f-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.595239 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9j62\" (UniqueName: \"kubernetes.io/projected/39e12872-2efe-4cec-82dd-a265f58da9ec-kube-api-access-s9j62\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.595251 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5pdb\" (UniqueName: \"kubernetes.io/projected/27c5a979-99f8-46a9-9a96-312c8c61186f-kube-api-access-v5pdb\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.595259 4998 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/27c5a979-99f8-46a9-9a96-312c8c61186f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.698413 4998 scope.go:117] "RemoveContainer" containerID="5254702d498c1d57539ad9a53df0e92c424d723090b3e059832ad451b99efb5f" Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.741200 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b6fcbcc87-jjnql"] Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.843817 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-594b65fc49-cgfb2"] Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.863432 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-594b65fc49-cgfb2"] Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.894823 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7786d8fd7-fmq4c"] Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.980740 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7786d8fd7-fmq4c"] Dec 03 16:22:26 crc kubenswrapper[4998]: I1203 16:22:26.999620 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-w7sx4"] Dec 03 16:22:27 crc kubenswrapper[4998]: I1203 16:22:27.042056 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56c4b567fc-78kqx"] Dec 03 16:22:27 crc kubenswrapper[4998]: I1203 16:22:27.111435 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:22:27 crc kubenswrapper[4998]: I1203 16:22:27.111506 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:22:27 crc kubenswrapper[4998]: I1203 16:22:27.111992 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:22:27 crc kubenswrapper[4998]: I1203 16:22:27.112717 4998 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4e819484c40f765e06b9e1c5a211c516f5099460ab695ea890ca1412f47f0680"} pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 16:22:27 crc kubenswrapper[4998]: I1203 16:22:27.113474 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" containerID="cri-o://4e819484c40f765e06b9e1c5a211c516f5099460ab695ea890ca1412f47f0680" gracePeriod=600 Dec 03 16:22:27 crc kubenswrapper[4998]: W1203 16:22:27.292223 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c6b2505_697f_4935_af28_88a4ae94524f.slice/crio-73f392ef5b749f98e6bee415211b68f67336f6059571034e2d1c4e63802db5e6 WatchSource:0}: Error finding container 73f392ef5b749f98e6bee415211b68f67336f6059571034e2d1c4e63802db5e6: Status 404 returned error can't find the container with id 73f392ef5b749f98e6bee415211b68f67336f6059571034e2d1c4e63802db5e6 Dec 03 16:22:27 crc kubenswrapper[4998]: W1203 16:22:27.301655 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podebaa6f1f_170e_41a4_ad4d_332108a0233b.slice/crio-e581127123aaf82f357938e0a52dbb462957c16da4440073caa2da5251413d2a WatchSource:0}: Error finding container e581127123aaf82f357938e0a52dbb462957c16da4440073caa2da5251413d2a: Status 404 returned error can't find the container with id e581127123aaf82f357938e0a52dbb462957c16da4440073caa2da5251413d2a Dec 03 16:22:27 crc kubenswrapper[4998]: W1203 16:22:27.303278 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddfe81aba_0bca_45d3_96f4_fd6ee0fe7580.slice/crio-4320eb91e2b02c39a7b211e817059f54a8638a03dfd535b2ffdc0616f07abce3 WatchSource:0}: Error finding container 4320eb91e2b02c39a7b211e817059f54a8638a03dfd535b2ffdc0616f07abce3: Status 404 returned error can't find the container with id 4320eb91e2b02c39a7b211e817059f54a8638a03dfd535b2ffdc0616f07abce3 Dec 03 16:22:27 crc kubenswrapper[4998]: I1203 16:22:27.309279 4998 scope.go:117] "RemoveContainer" containerID="1cdb1825530aab9db06bbdc28a5a98ec78ab0055779a12059dbf2c4a0e0e57d2" Dec 03 16:22:27 crc kubenswrapper[4998]: E1203 16:22:27.309617 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cdb1825530aab9db06bbdc28a5a98ec78ab0055779a12059dbf2c4a0e0e57d2\": container with ID starting with 1cdb1825530aab9db06bbdc28a5a98ec78ab0055779a12059dbf2c4a0e0e57d2 not found: ID does not exist" containerID="1cdb1825530aab9db06bbdc28a5a98ec78ab0055779a12059dbf2c4a0e0e57d2" Dec 03 16:22:27 crc kubenswrapper[4998]: I1203 16:22:27.309648 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cdb1825530aab9db06bbdc28a5a98ec78ab0055779a12059dbf2c4a0e0e57d2"} err="failed to get container status \"1cdb1825530aab9db06bbdc28a5a98ec78ab0055779a12059dbf2c4a0e0e57d2\": rpc error: code = NotFound desc = could not find container \"1cdb1825530aab9db06bbdc28a5a98ec78ab0055779a12059dbf2c4a0e0e57d2\": container with ID starting with 1cdb1825530aab9db06bbdc28a5a98ec78ab0055779a12059dbf2c4a0e0e57d2 not found: ID does not exist" Dec 03 16:22:27 crc kubenswrapper[4998]: I1203 16:22:27.309671 4998 scope.go:117] "RemoveContainer" containerID="5254702d498c1d57539ad9a53df0e92c424d723090b3e059832ad451b99efb5f" Dec 03 16:22:27 crc kubenswrapper[4998]: E1203 16:22:27.310044 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5254702d498c1d57539ad9a53df0e92c424d723090b3e059832ad451b99efb5f\": container with ID starting with 5254702d498c1d57539ad9a53df0e92c424d723090b3e059832ad451b99efb5f not found: ID does not exist" containerID="5254702d498c1d57539ad9a53df0e92c424d723090b3e059832ad451b99efb5f" Dec 03 16:22:27 crc kubenswrapper[4998]: I1203 16:22:27.310065 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5254702d498c1d57539ad9a53df0e92c424d723090b3e059832ad451b99efb5f"} err="failed to get container status \"5254702d498c1d57539ad9a53df0e92c424d723090b3e059832ad451b99efb5f\": rpc error: code = NotFound desc = could not find container \"5254702d498c1d57539ad9a53df0e92c424d723090b3e059832ad451b99efb5f\": container with ID starting with 5254702d498c1d57539ad9a53df0e92c424d723090b3e059832ad451b99efb5f not found: ID does not exist" Dec 03 16:22:27 crc kubenswrapper[4998]: I1203 16:22:27.310082 4998 scope.go:117] "RemoveContainer" containerID="b9ec91c1bee81c5e74567c92b09c45e79715222f9b173a4083c26528bfd1ec04" Dec 03 16:22:27 crc kubenswrapper[4998]: I1203 16:22:27.450416 4998 scope.go:117] "RemoveContainer" containerID="d51f2297cf0d8f82d6a221d7ec19a91a1af35082e627ac9c26314008339d3036" Dec 03 16:22:27 crc kubenswrapper[4998]: I1203 16:22:27.489072 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6fcbcc87-jjnql" event={"ID":"4c6b2505-697f-4935-af28-88a4ae94524f","Type":"ContainerStarted","Data":"73f392ef5b749f98e6bee415211b68f67336f6059571034e2d1c4e63802db5e6"} Dec 03 16:22:27 crc kubenswrapper[4998]: I1203 16:22:27.492278 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-w7sx4" event={"ID":"ebaa6f1f-170e-41a4-ad4d-332108a0233b","Type":"ContainerStarted","Data":"e581127123aaf82f357938e0a52dbb462957c16da4440073caa2da5251413d2a"} Dec 03 16:22:27 crc kubenswrapper[4998]: I1203 16:22:27.495387 4998 generic.go:334] "Generic (PLEG): container finished" podID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerID="4e819484c40f765e06b9e1c5a211c516f5099460ab695ea890ca1412f47f0680" exitCode=0 Dec 03 16:22:27 crc kubenswrapper[4998]: I1203 16:22:27.495438 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerDied","Data":"4e819484c40f765e06b9e1c5a211c516f5099460ab695ea890ca1412f47f0680"} Dec 03 16:22:27 crc kubenswrapper[4998]: I1203 16:22:27.495469 4998 scope.go:117] "RemoveContainer" containerID="5b97e0b62564fae4effbf36e0feb1d3ce5f2c4a4d5ac7c086146fd76fc8ee224" Dec 03 16:22:27 crc kubenswrapper[4998]: I1203 16:22:27.500310 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" event={"ID":"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580","Type":"ContainerStarted","Data":"4320eb91e2b02c39a7b211e817059f54a8638a03dfd535b2ffdc0616f07abce3"} Dec 03 16:22:27 crc kubenswrapper[4998]: I1203 16:22:27.687095 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27c5a979-99f8-46a9-9a96-312c8c61186f" path="/var/lib/kubelet/pods/27c5a979-99f8-46a9-9a96-312c8c61186f/volumes" Dec 03 16:22:27 crc kubenswrapper[4998]: I1203 16:22:27.687834 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39e12872-2efe-4cec-82dd-a265f58da9ec" path="/var/lib/kubelet/pods/39e12872-2efe-4cec-82dd-a265f58da9ec/volumes" Dec 03 16:22:28 crc kubenswrapper[4998]: I1203 16:22:28.510697 4998 generic.go:334] "Generic (PLEG): container finished" podID="dfe81aba-0bca-45d3-96f4-fd6ee0fe7580" containerID="b7922d628f552d552b3534838a2a83c5ec05bc127893b947c0d2d23f6ca59187" exitCode=0 Dec 03 16:22:28 crc kubenswrapper[4998]: I1203 16:22:28.511331 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" event={"ID":"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580","Type":"ContainerDied","Data":"b7922d628f552d552b3534838a2a83c5ec05bc127893b947c0d2d23f6ca59187"} Dec 03 16:22:28 crc kubenswrapper[4998]: I1203 16:22:28.515181 4998 generic.go:334] "Generic (PLEG): container finished" podID="4c6b2505-697f-4935-af28-88a4ae94524f" containerID="47cdf182c8e1aec57071b2479434e6a1905d6565e7dc73e1564b3e2f9b833d55" exitCode=0 Dec 03 16:22:28 crc kubenswrapper[4998]: I1203 16:22:28.515261 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6fcbcc87-jjnql" event={"ID":"4c6b2505-697f-4935-af28-88a4ae94524f","Type":"ContainerDied","Data":"47cdf182c8e1aec57071b2479434e6a1905d6565e7dc73e1564b3e2f9b833d55"} Dec 03 16:22:28 crc kubenswrapper[4998]: I1203 16:22:28.519015 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5hxf2" event={"ID":"832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5","Type":"ContainerStarted","Data":"e244bb6795e9b2fb01bd1238cfacc397b4f27cc443422dbda834b17f2e573073"} Dec 03 16:22:28 crc kubenswrapper[4998]: I1203 16:22:28.520424 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf","Type":"ContainerStarted","Data":"b215993216cc58899b597c45917e79e3e2bad83f6bf1e8114519bf2dfd1b4128"} Dec 03 16:22:28 crc kubenswrapper[4998]: I1203 16:22:28.522743 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"dc47f739351bd606edc721ca357f0c89b466b48ca6131ea4bbb968b4e4cc02c4"} Dec 03 16:22:29 crc kubenswrapper[4998]: I1203 16:22:29.485182 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 03 16:22:30 crc kubenswrapper[4998]: I1203 16:22:30.551795 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"893cf12f-77a9-463f-8dd4-1b3f05b2c9bf","Type":"ContainerStarted","Data":"cb8cd4d9e97d37908127a382255c1465a7d845c90f59caa892b9057ddab5b402"} Dec 03 16:22:30 crc kubenswrapper[4998]: I1203 16:22:30.553678 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-w7sx4" event={"ID":"ebaa6f1f-170e-41a4-ad4d-332108a0233b","Type":"ContainerStarted","Data":"817ad9b8c1ce1b1cb7ff1324bb4dd221deaa9f34a66e5a22ab2d547f321254d9"} Dec 03 16:22:30 crc kubenswrapper[4998]: I1203 16:22:30.555330 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f9abb1de-7121-4de9-833a-9f1ab889b8ff","Type":"ContainerStarted","Data":"3a0b0d4f0a8d7f25f121dff94ba7f15e27e33f0d83a805c785800edd09699873"} Dec 03 16:22:30 crc kubenswrapper[4998]: I1203 16:22:30.555459 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 16:22:30 crc kubenswrapper[4998]: I1203 16:22:30.557087 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" event={"ID":"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580","Type":"ContainerStarted","Data":"a51dec14a58fa537bd3e31f8c8880f0ddf874a18fa4fcb92cda4d91716e7b04d"} Dec 03 16:22:30 crc kubenswrapper[4998]: I1203 16:22:30.557224 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" Dec 03 16:22:30 crc kubenswrapper[4998]: I1203 16:22:30.558666 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6fcbcc87-jjnql" event={"ID":"4c6b2505-697f-4935-af28-88a4ae94524f","Type":"ContainerStarted","Data":"18584dc50961546fa86c00b8e71d0469c200cf58ca1878e37a80d8b933964890"} Dec 03 16:22:30 crc kubenswrapper[4998]: I1203 16:22:30.558745 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b6fcbcc87-jjnql" Dec 03 16:22:30 crc kubenswrapper[4998]: I1203 16:22:30.560347 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"37aa6aee-15f8-4e23-a18c-216b85e605cf","Type":"ContainerStarted","Data":"0e061a4eafd36ccf8b3584fecef4e537e351d3513d41c019997e14ad4aa9f458"} Dec 03 16:22:30 crc kubenswrapper[4998]: I1203 16:22:30.562352 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-5hxf2" event={"ID":"832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5","Type":"ContainerStarted","Data":"cab0568d2dd1f5e4be11c5f01b881d8aaaa41452298b94caf773708d5ec489f4"} Dec 03 16:22:30 crc kubenswrapper[4998]: I1203 16:22:30.562495 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-5hxf2" Dec 03 16:22:30 crc kubenswrapper[4998]: I1203 16:22:30.576444 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=18.866631919 podStartE2EDuration="31.576418178s" podCreationTimestamp="2025-12-03 16:21:59 +0000 UTC" firstStartedPulling="2025-12-03 16:22:17.250058696 +0000 UTC m=+1115.861758919" lastFinishedPulling="2025-12-03 16:22:29.959844955 +0000 UTC m=+1128.571545178" observedRunningTime="2025-12-03 16:22:30.568498449 +0000 UTC m=+1129.180198672" watchObservedRunningTime="2025-12-03 16:22:30.576418178 +0000 UTC m=+1129.188118401" Dec 03 16:22:30 crc kubenswrapper[4998]: I1203 16:22:30.592096 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=11.168695663 podStartE2EDuration="28.592077443s" podCreationTimestamp="2025-12-03 16:22:02 +0000 UTC" firstStartedPulling="2025-12-03 16:22:12.537016029 +0000 UTC m=+1111.148716242" lastFinishedPulling="2025-12-03 16:22:29.960397799 +0000 UTC m=+1128.572098022" observedRunningTime="2025-12-03 16:22:30.59038687 +0000 UTC m=+1129.202087093" watchObservedRunningTime="2025-12-03 16:22:30.592077443 +0000 UTC m=+1129.203777666" Dec 03 16:22:30 crc kubenswrapper[4998]: I1203 16:22:30.622398 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" podStartSLOduration=7.622384166 podStartE2EDuration="7.622384166s" podCreationTimestamp="2025-12-03 16:22:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:22:30.618930929 +0000 UTC m=+1129.230631152" watchObservedRunningTime="2025-12-03 16:22:30.622384166 +0000 UTC m=+1129.234084389" Dec 03 16:22:30 crc kubenswrapper[4998]: I1203 16:22:30.643178 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-w7sx4" podStartSLOduration=4.994610265 podStartE2EDuration="7.64315822s" podCreationTimestamp="2025-12-03 16:22:23 +0000 UTC" firstStartedPulling="2025-12-03 16:22:27.310234613 +0000 UTC m=+1125.921934836" lastFinishedPulling="2025-12-03 16:22:29.958782568 +0000 UTC m=+1128.570482791" observedRunningTime="2025-12-03 16:22:30.637056146 +0000 UTC m=+1129.248756369" watchObservedRunningTime="2025-12-03 16:22:30.64315822 +0000 UTC m=+1129.254858433" Dec 03 16:22:30 crc kubenswrapper[4998]: I1203 16:22:30.670448 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=16.987491879 podStartE2EDuration="34.670430487s" podCreationTimestamp="2025-12-03 16:21:56 +0000 UTC" firstStartedPulling="2025-12-03 16:22:12.277294036 +0000 UTC m=+1110.888994259" lastFinishedPulling="2025-12-03 16:22:29.960232644 +0000 UTC m=+1128.571932867" observedRunningTime="2025-12-03 16:22:30.657978403 +0000 UTC m=+1129.269678626" watchObservedRunningTime="2025-12-03 16:22:30.670430487 +0000 UTC m=+1129.282130710" Dec 03 16:22:30 crc kubenswrapper[4998]: I1203 16:22:30.690418 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-5hxf2" podStartSLOduration=26.418853264 podStartE2EDuration="31.69039465s" podCreationTimestamp="2025-12-03 16:21:59 +0000 UTC" firstStartedPulling="2025-12-03 16:22:17.250028865 +0000 UTC m=+1115.861729108" lastFinishedPulling="2025-12-03 16:22:22.521570241 +0000 UTC m=+1121.133270494" observedRunningTime="2025-12-03 16:22:30.684681176 +0000 UTC m=+1129.296381399" watchObservedRunningTime="2025-12-03 16:22:30.69039465 +0000 UTC m=+1129.302094873" Dec 03 16:22:30 crc kubenswrapper[4998]: I1203 16:22:30.707783 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b6fcbcc87-jjnql" podStartSLOduration=7.707745757 podStartE2EDuration="7.707745757s" podCreationTimestamp="2025-12-03 16:22:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:22:30.703243953 +0000 UTC m=+1129.314944176" watchObservedRunningTime="2025-12-03 16:22:30.707745757 +0000 UTC m=+1129.319445980" Dec 03 16:22:30 crc kubenswrapper[4998]: I1203 16:22:30.883350 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:30 crc kubenswrapper[4998]: I1203 16:22:30.939911 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:31 crc kubenswrapper[4998]: I1203 16:22:31.571058 4998 generic.go:334] "Generic (PLEG): container finished" podID="8ea3f8b4-86fc-4888-b9c0-36859ab25724" containerID="d65ca50f1c810cea0895251df4ced37c187aab806e40c39e892ccc774306b756" exitCode=0 Dec 03 16:22:31 crc kubenswrapper[4998]: I1203 16:22:31.571101 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8ea3f8b4-86fc-4888-b9c0-36859ab25724","Type":"ContainerDied","Data":"d65ca50f1c810cea0895251df4ced37c187aab806e40c39e892ccc774306b756"} Dec 03 16:22:31 crc kubenswrapper[4998]: I1203 16:22:31.574589 4998 generic.go:334] "Generic (PLEG): container finished" podID="1b2c3df0-703c-438e-9887-cf29844b9eb7" containerID="6e37399ddc89c2d43698c7ae97f2b25c082586d474cceb3b7c586516d81af9fd" exitCode=0 Dec 03 16:22:31 crc kubenswrapper[4998]: I1203 16:22:31.574727 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1b2c3df0-703c-438e-9887-cf29844b9eb7","Type":"ContainerDied","Data":"6e37399ddc89c2d43698c7ae97f2b25c082586d474cceb3b7c586516d81af9fd"} Dec 03 16:22:31 crc kubenswrapper[4998]: I1203 16:22:31.576180 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-5hxf2" Dec 03 16:22:31 crc kubenswrapper[4998]: I1203 16:22:31.576493 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:31 crc kubenswrapper[4998]: I1203 16:22:31.631681 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 03 16:22:32 crc kubenswrapper[4998]: I1203 16:22:32.586155 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8ea3f8b4-86fc-4888-b9c0-36859ab25724","Type":"ContainerStarted","Data":"c3d29a098e149ff2e56f0579f2fc7f989170b9c8245e87ef03dfd98377648c19"} Dec 03 16:22:32 crc kubenswrapper[4998]: I1203 16:22:32.589417 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1b2c3df0-703c-438e-9887-cf29844b9eb7","Type":"ContainerStarted","Data":"b7a87b0325114612b243211a832913217372ad997148e5bbeaabec50b1fbf2cb"} Dec 03 16:22:32 crc kubenswrapper[4998]: I1203 16:22:32.615868 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=31.279219815 podStartE2EDuration="41.615850657s" podCreationTimestamp="2025-12-03 16:21:51 +0000 UTC" firstStartedPulling="2025-12-03 16:22:12.188243552 +0000 UTC m=+1110.799943775" lastFinishedPulling="2025-12-03 16:22:22.524874364 +0000 UTC m=+1121.136574617" observedRunningTime="2025-12-03 16:22:32.614571035 +0000 UTC m=+1131.226271288" watchObservedRunningTime="2025-12-03 16:22:32.615850657 +0000 UTC m=+1131.227550870" Dec 03 16:22:32 crc kubenswrapper[4998]: I1203 16:22:32.644531 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=31.585134689 podStartE2EDuration="39.64451293s" podCreationTimestamp="2025-12-03 16:21:53 +0000 UTC" firstStartedPulling="2025-12-03 16:22:12.195124436 +0000 UTC m=+1110.806824669" lastFinishedPulling="2025-12-03 16:22:20.254502687 +0000 UTC m=+1118.866202910" observedRunningTime="2025-12-03 16:22:32.638352494 +0000 UTC m=+1131.250052717" watchObservedRunningTime="2025-12-03 16:22:32.64451293 +0000 UTC m=+1131.256213153" Dec 03 16:22:33 crc kubenswrapper[4998]: I1203 16:22:33.127387 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 03 16:22:33 crc kubenswrapper[4998]: I1203 16:22:33.127471 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 03 16:22:33 crc kubenswrapper[4998]: I1203 16:22:33.471472 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:33 crc kubenswrapper[4998]: I1203 16:22:33.526637 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:33 crc kubenswrapper[4998]: I1203 16:22:33.599679 4998 generic.go:334] "Generic (PLEG): container finished" podID="92d2d9f9-b0c3-40f8-981a-41ec47f030e1" containerID="174a248e64d7d0441f1074334990836bdc3697b8f54e27015258af83a045fcd0" exitCode=0 Dec 03 16:22:33 crc kubenswrapper[4998]: I1203 16:22:33.599799 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92d2d9f9-b0c3-40f8-981a-41ec47f030e1","Type":"ContainerDied","Data":"174a248e64d7d0441f1074334990836bdc3697b8f54e27015258af83a045fcd0"} Dec 03 16:22:33 crc kubenswrapper[4998]: I1203 16:22:33.600368 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:33 crc kubenswrapper[4998]: I1203 16:22:33.664392 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 03 16:22:33 crc kubenswrapper[4998]: I1203 16:22:33.931669 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 03 16:22:33 crc kubenswrapper[4998]: E1203 16:22:33.932392 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39e12872-2efe-4cec-82dd-a265f58da9ec" containerName="init" Dec 03 16:22:33 crc kubenswrapper[4998]: I1203 16:22:33.932408 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="39e12872-2efe-4cec-82dd-a265f58da9ec" containerName="init" Dec 03 16:22:33 crc kubenswrapper[4998]: E1203 16:22:33.932417 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27c5a979-99f8-46a9-9a96-312c8c61186f" containerName="init" Dec 03 16:22:33 crc kubenswrapper[4998]: I1203 16:22:33.932423 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="27c5a979-99f8-46a9-9a96-312c8c61186f" containerName="init" Dec 03 16:22:33 crc kubenswrapper[4998]: E1203 16:22:33.932441 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27c5a979-99f8-46a9-9a96-312c8c61186f" containerName="dnsmasq-dns" Dec 03 16:22:33 crc kubenswrapper[4998]: I1203 16:22:33.932450 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="27c5a979-99f8-46a9-9a96-312c8c61186f" containerName="dnsmasq-dns" Dec 03 16:22:33 crc kubenswrapper[4998]: E1203 16:22:33.932469 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39e12872-2efe-4cec-82dd-a265f58da9ec" containerName="dnsmasq-dns" Dec 03 16:22:33 crc kubenswrapper[4998]: I1203 16:22:33.932475 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="39e12872-2efe-4cec-82dd-a265f58da9ec" containerName="dnsmasq-dns" Dec 03 16:22:33 crc kubenswrapper[4998]: I1203 16:22:33.932634 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="27c5a979-99f8-46a9-9a96-312c8c61186f" containerName="dnsmasq-dns" Dec 03 16:22:33 crc kubenswrapper[4998]: I1203 16:22:33.932651 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="39e12872-2efe-4cec-82dd-a265f58da9ec" containerName="dnsmasq-dns" Dec 03 16:22:33 crc kubenswrapper[4998]: I1203 16:22:33.937808 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 16:22:33 crc kubenswrapper[4998]: I1203 16:22:33.940071 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-hvkdx" Dec 03 16:22:33 crc kubenswrapper[4998]: I1203 16:22:33.940291 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 03 16:22:33 crc kubenswrapper[4998]: I1203 16:22:33.940442 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 03 16:22:33 crc kubenswrapper[4998]: I1203 16:22:33.944194 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 03 16:22:33 crc kubenswrapper[4998]: I1203 16:22:33.961131 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 16:22:34 crc kubenswrapper[4998]: I1203 16:22:34.060609 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvv6p\" (UniqueName: \"kubernetes.io/projected/4e93efdf-68ff-4d81-b6b3-447a482e74ff-kube-api-access-kvv6p\") pod \"ovn-northd-0\" (UID: \"4e93efdf-68ff-4d81-b6b3-447a482e74ff\") " pod="openstack/ovn-northd-0" Dec 03 16:22:34 crc kubenswrapper[4998]: I1203 16:22:34.060653 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e93efdf-68ff-4d81-b6b3-447a482e74ff-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"4e93efdf-68ff-4d81-b6b3-447a482e74ff\") " pod="openstack/ovn-northd-0" Dec 03 16:22:34 crc kubenswrapper[4998]: I1203 16:22:34.060694 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e93efdf-68ff-4d81-b6b3-447a482e74ff-config\") pod \"ovn-northd-0\" (UID: \"4e93efdf-68ff-4d81-b6b3-447a482e74ff\") " pod="openstack/ovn-northd-0" Dec 03 16:22:34 crc kubenswrapper[4998]: I1203 16:22:34.060720 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4e93efdf-68ff-4d81-b6b3-447a482e74ff-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4e93efdf-68ff-4d81-b6b3-447a482e74ff\") " pod="openstack/ovn-northd-0" Dec 03 16:22:34 crc kubenswrapper[4998]: I1203 16:22:34.060744 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e93efdf-68ff-4d81-b6b3-447a482e74ff-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"4e93efdf-68ff-4d81-b6b3-447a482e74ff\") " pod="openstack/ovn-northd-0" Dec 03 16:22:34 crc kubenswrapper[4998]: I1203 16:22:34.060779 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e93efdf-68ff-4d81-b6b3-447a482e74ff-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4e93efdf-68ff-4d81-b6b3-447a482e74ff\") " pod="openstack/ovn-northd-0" Dec 03 16:22:34 crc kubenswrapper[4998]: I1203 16:22:34.060813 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e93efdf-68ff-4d81-b6b3-447a482e74ff-scripts\") pod \"ovn-northd-0\" (UID: \"4e93efdf-68ff-4d81-b6b3-447a482e74ff\") " pod="openstack/ovn-northd-0" Dec 03 16:22:34 crc kubenswrapper[4998]: I1203 16:22:34.161799 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e93efdf-68ff-4d81-b6b3-447a482e74ff-config\") pod \"ovn-northd-0\" (UID: \"4e93efdf-68ff-4d81-b6b3-447a482e74ff\") " pod="openstack/ovn-northd-0" Dec 03 16:22:34 crc kubenswrapper[4998]: I1203 16:22:34.161868 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4e93efdf-68ff-4d81-b6b3-447a482e74ff-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4e93efdf-68ff-4d81-b6b3-447a482e74ff\") " pod="openstack/ovn-northd-0" Dec 03 16:22:34 crc kubenswrapper[4998]: I1203 16:22:34.161910 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e93efdf-68ff-4d81-b6b3-447a482e74ff-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"4e93efdf-68ff-4d81-b6b3-447a482e74ff\") " pod="openstack/ovn-northd-0" Dec 03 16:22:34 crc kubenswrapper[4998]: I1203 16:22:34.161945 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e93efdf-68ff-4d81-b6b3-447a482e74ff-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4e93efdf-68ff-4d81-b6b3-447a482e74ff\") " pod="openstack/ovn-northd-0" Dec 03 16:22:34 crc kubenswrapper[4998]: I1203 16:22:34.162002 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e93efdf-68ff-4d81-b6b3-447a482e74ff-scripts\") pod \"ovn-northd-0\" (UID: \"4e93efdf-68ff-4d81-b6b3-447a482e74ff\") " pod="openstack/ovn-northd-0" Dec 03 16:22:34 crc kubenswrapper[4998]: I1203 16:22:34.162089 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvv6p\" (UniqueName: \"kubernetes.io/projected/4e93efdf-68ff-4d81-b6b3-447a482e74ff-kube-api-access-kvv6p\") pod \"ovn-northd-0\" (UID: \"4e93efdf-68ff-4d81-b6b3-447a482e74ff\") " pod="openstack/ovn-northd-0" Dec 03 16:22:34 crc kubenswrapper[4998]: I1203 16:22:34.162119 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e93efdf-68ff-4d81-b6b3-447a482e74ff-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"4e93efdf-68ff-4d81-b6b3-447a482e74ff\") " pod="openstack/ovn-northd-0" Dec 03 16:22:34 crc kubenswrapper[4998]: I1203 16:22:34.162912 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4e93efdf-68ff-4d81-b6b3-447a482e74ff-scripts\") pod \"ovn-northd-0\" (UID: \"4e93efdf-68ff-4d81-b6b3-447a482e74ff\") " pod="openstack/ovn-northd-0" Dec 03 16:22:34 crc kubenswrapper[4998]: I1203 16:22:34.162961 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e93efdf-68ff-4d81-b6b3-447a482e74ff-config\") pod \"ovn-northd-0\" (UID: \"4e93efdf-68ff-4d81-b6b3-447a482e74ff\") " pod="openstack/ovn-northd-0" Dec 03 16:22:34 crc kubenswrapper[4998]: I1203 16:22:34.163247 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4e93efdf-68ff-4d81-b6b3-447a482e74ff-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4e93efdf-68ff-4d81-b6b3-447a482e74ff\") " pod="openstack/ovn-northd-0" Dec 03 16:22:34 crc kubenswrapper[4998]: I1203 16:22:34.168572 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e93efdf-68ff-4d81-b6b3-447a482e74ff-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"4e93efdf-68ff-4d81-b6b3-447a482e74ff\") " pod="openstack/ovn-northd-0" Dec 03 16:22:34 crc kubenswrapper[4998]: I1203 16:22:34.168827 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e93efdf-68ff-4d81-b6b3-447a482e74ff-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"4e93efdf-68ff-4d81-b6b3-447a482e74ff\") " pod="openstack/ovn-northd-0" Dec 03 16:22:34 crc kubenswrapper[4998]: I1203 16:22:34.169031 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e93efdf-68ff-4d81-b6b3-447a482e74ff-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4e93efdf-68ff-4d81-b6b3-447a482e74ff\") " pod="openstack/ovn-northd-0" Dec 03 16:22:34 crc kubenswrapper[4998]: I1203 16:22:34.183817 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvv6p\" (UniqueName: \"kubernetes.io/projected/4e93efdf-68ff-4d81-b6b3-447a482e74ff-kube-api-access-kvv6p\") pod \"ovn-northd-0\" (UID: \"4e93efdf-68ff-4d81-b6b3-447a482e74ff\") " pod="openstack/ovn-northd-0" Dec 03 16:22:34 crc kubenswrapper[4998]: I1203 16:22:34.263357 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 16:22:34 crc kubenswrapper[4998]: I1203 16:22:34.420172 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 03 16:22:34 crc kubenswrapper[4998]: I1203 16:22:34.420577 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 03 16:22:34 crc kubenswrapper[4998]: I1203 16:22:34.725534 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 16:22:34 crc kubenswrapper[4998]: W1203 16:22:34.734728 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e93efdf_68ff_4d81_b6b3_447a482e74ff.slice/crio-53e46b3c1774ef5ba20b1118e6379fa5fc5647597f5da3cc138b1e8a2449211a WatchSource:0}: Error finding container 53e46b3c1774ef5ba20b1118e6379fa5fc5647597f5da3cc138b1e8a2449211a: Status 404 returned error can't find the container with id 53e46b3c1774ef5ba20b1118e6379fa5fc5647597f5da3cc138b1e8a2449211a Dec 03 16:22:35 crc kubenswrapper[4998]: I1203 16:22:35.640442 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4e93efdf-68ff-4d81-b6b3-447a482e74ff","Type":"ContainerStarted","Data":"53e46b3c1774ef5ba20b1118e6379fa5fc5647597f5da3cc138b1e8a2449211a"} Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.644128 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.678510 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4e93efdf-68ff-4d81-b6b3-447a482e74ff","Type":"ContainerStarted","Data":"7f082bbae48108a4910f7d62b97428c6a61d907c4962e87df3c5640c5cd3b72a"} Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.678566 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4e93efdf-68ff-4d81-b6b3-447a482e74ff","Type":"ContainerStarted","Data":"fc818b5d71463de3ca8bb74121fabe2cf7e8fabaf3c057df691ef85c226b5371"} Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.679427 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.683084 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56c4b567fc-78kqx"] Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.683417 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" podUID="dfe81aba-0bca-45d3-96f4-fd6ee0fe7580" containerName="dnsmasq-dns" containerID="cri-o://a51dec14a58fa537bd3e31f8c8880f0ddf874a18fa4fcb92cda4d91716e7b04d" gracePeriod=10 Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.689884 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.719303 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55ddfd5dfc-9clrt"] Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.721007 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.723319 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.07602141 podStartE2EDuration="3.723303517s" podCreationTimestamp="2025-12-03 16:22:33 +0000 UTC" firstStartedPulling="2025-12-03 16:22:34.737286904 +0000 UTC m=+1133.348987127" lastFinishedPulling="2025-12-03 16:22:35.384569011 +0000 UTC m=+1133.996269234" observedRunningTime="2025-12-03 16:22:36.714949537 +0000 UTC m=+1135.326649760" watchObservedRunningTime="2025-12-03 16:22:36.723303517 +0000 UTC m=+1135.335003740" Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.745204 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55ddfd5dfc-9clrt"] Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.806635 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e266f963-33c9-48dc-bb10-70bd5c700d69-ovsdbserver-sb\") pod \"dnsmasq-dns-55ddfd5dfc-9clrt\" (UID: \"e266f963-33c9-48dc-bb10-70bd5c700d69\") " pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.806685 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnvj4\" (UniqueName: \"kubernetes.io/projected/e266f963-33c9-48dc-bb10-70bd5c700d69-kube-api-access-xnvj4\") pod \"dnsmasq-dns-55ddfd5dfc-9clrt\" (UID: \"e266f963-33c9-48dc-bb10-70bd5c700d69\") " pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.806721 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e266f963-33c9-48dc-bb10-70bd5c700d69-config\") pod \"dnsmasq-dns-55ddfd5dfc-9clrt\" (UID: \"e266f963-33c9-48dc-bb10-70bd5c700d69\") " pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.806818 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e266f963-33c9-48dc-bb10-70bd5c700d69-dns-svc\") pod \"dnsmasq-dns-55ddfd5dfc-9clrt\" (UID: \"e266f963-33c9-48dc-bb10-70bd5c700d69\") " pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.806861 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e266f963-33c9-48dc-bb10-70bd5c700d69-ovsdbserver-nb\") pod \"dnsmasq-dns-55ddfd5dfc-9clrt\" (UID: \"e266f963-33c9-48dc-bb10-70bd5c700d69\") " pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.907734 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e266f963-33c9-48dc-bb10-70bd5c700d69-ovsdbserver-nb\") pod \"dnsmasq-dns-55ddfd5dfc-9clrt\" (UID: \"e266f963-33c9-48dc-bb10-70bd5c700d69\") " pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.907824 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e266f963-33c9-48dc-bb10-70bd5c700d69-ovsdbserver-sb\") pod \"dnsmasq-dns-55ddfd5dfc-9clrt\" (UID: \"e266f963-33c9-48dc-bb10-70bd5c700d69\") " pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.907850 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnvj4\" (UniqueName: \"kubernetes.io/projected/e266f963-33c9-48dc-bb10-70bd5c700d69-kube-api-access-xnvj4\") pod \"dnsmasq-dns-55ddfd5dfc-9clrt\" (UID: \"e266f963-33c9-48dc-bb10-70bd5c700d69\") " pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.907883 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e266f963-33c9-48dc-bb10-70bd5c700d69-config\") pod \"dnsmasq-dns-55ddfd5dfc-9clrt\" (UID: \"e266f963-33c9-48dc-bb10-70bd5c700d69\") " pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.907940 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e266f963-33c9-48dc-bb10-70bd5c700d69-dns-svc\") pod \"dnsmasq-dns-55ddfd5dfc-9clrt\" (UID: \"e266f963-33c9-48dc-bb10-70bd5c700d69\") " pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.908818 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e266f963-33c9-48dc-bb10-70bd5c700d69-ovsdbserver-nb\") pod \"dnsmasq-dns-55ddfd5dfc-9clrt\" (UID: \"e266f963-33c9-48dc-bb10-70bd5c700d69\") " pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.908822 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e266f963-33c9-48dc-bb10-70bd5c700d69-dns-svc\") pod \"dnsmasq-dns-55ddfd5dfc-9clrt\" (UID: \"e266f963-33c9-48dc-bb10-70bd5c700d69\") " pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.909903 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e266f963-33c9-48dc-bb10-70bd5c700d69-config\") pod \"dnsmasq-dns-55ddfd5dfc-9clrt\" (UID: \"e266f963-33c9-48dc-bb10-70bd5c700d69\") " pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.910204 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e266f963-33c9-48dc-bb10-70bd5c700d69-ovsdbserver-sb\") pod \"dnsmasq-dns-55ddfd5dfc-9clrt\" (UID: \"e266f963-33c9-48dc-bb10-70bd5c700d69\") " pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" Dec 03 16:22:36 crc kubenswrapper[4998]: I1203 16:22:36.946918 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnvj4\" (UniqueName: \"kubernetes.io/projected/e266f963-33c9-48dc-bb10-70bd5c700d69-kube-api-access-xnvj4\") pod \"dnsmasq-dns-55ddfd5dfc-9clrt\" (UID: \"e266f963-33c9-48dc-bb10-70bd5c700d69\") " pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.134270 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.255372 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.313157 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgf84\" (UniqueName: \"kubernetes.io/projected/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-kube-api-access-dgf84\") pod \"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580\" (UID: \"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580\") " Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.313268 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-ovsdbserver-nb\") pod \"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580\" (UID: \"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580\") " Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.313302 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-ovsdbserver-sb\") pod \"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580\" (UID: \"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580\") " Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.313352 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-config\") pod \"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580\" (UID: \"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580\") " Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.313437 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-dns-svc\") pod \"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580\" (UID: \"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580\") " Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.336448 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-kube-api-access-dgf84" (OuterVolumeSpecName: "kube-api-access-dgf84") pod "dfe81aba-0bca-45d3-96f4-fd6ee0fe7580" (UID: "dfe81aba-0bca-45d3-96f4-fd6ee0fe7580"). InnerVolumeSpecName "kube-api-access-dgf84". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.372694 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "dfe81aba-0bca-45d3-96f4-fd6ee0fe7580" (UID: "dfe81aba-0bca-45d3-96f4-fd6ee0fe7580"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.377825 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-config" (OuterVolumeSpecName: "config") pod "dfe81aba-0bca-45d3-96f4-fd6ee0fe7580" (UID: "dfe81aba-0bca-45d3-96f4-fd6ee0fe7580"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.389143 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "dfe81aba-0bca-45d3-96f4-fd6ee0fe7580" (UID: "dfe81aba-0bca-45d3-96f4-fd6ee0fe7580"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.390220 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dfe81aba-0bca-45d3-96f4-fd6ee0fe7580" (UID: "dfe81aba-0bca-45d3-96f4-fd6ee0fe7580"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.419010 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.419058 4998 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.419073 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgf84\" (UniqueName: \"kubernetes.io/projected/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-kube-api-access-dgf84\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.419086 4998 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.419098 4998 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.709275 4998 generic.go:334] "Generic (PLEG): container finished" podID="dfe81aba-0bca-45d3-96f4-fd6ee0fe7580" containerID="a51dec14a58fa537bd3e31f8c8880f0ddf874a18fa4fcb92cda4d91716e7b04d" exitCode=0 Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.710859 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.712656 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" event={"ID":"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580","Type":"ContainerDied","Data":"a51dec14a58fa537bd3e31f8c8880f0ddf874a18fa4fcb92cda4d91716e7b04d"} Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.712712 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56c4b567fc-78kqx" event={"ID":"dfe81aba-0bca-45d3-96f4-fd6ee0fe7580","Type":"ContainerDied","Data":"4320eb91e2b02c39a7b211e817059f54a8638a03dfd535b2ffdc0616f07abce3"} Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.712745 4998 scope.go:117] "RemoveContainer" containerID="a51dec14a58fa537bd3e31f8c8880f0ddf874a18fa4fcb92cda4d91716e7b04d" Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.743466 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55ddfd5dfc-9clrt"] Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.747897 4998 scope.go:117] "RemoveContainer" containerID="b7922d628f552d552b3534838a2a83c5ec05bc127893b947c0d2d23f6ca59187" Dec 03 16:22:37 crc kubenswrapper[4998]: W1203 16:22:37.755920 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode266f963_33c9_48dc_bb10_70bd5c700d69.slice/crio-7e6b8fcb2c28749e83376531d1ca7c381481f59b2a757cb53d68873facac9224 WatchSource:0}: Error finding container 7e6b8fcb2c28749e83376531d1ca7c381481f59b2a757cb53d68873facac9224: Status 404 returned error can't find the container with id 7e6b8fcb2c28749e83376531d1ca7c381481f59b2a757cb53d68873facac9224 Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.774425 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56c4b567fc-78kqx"] Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.790444 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56c4b567fc-78kqx"] Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.855025 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 03 16:22:37 crc kubenswrapper[4998]: E1203 16:22:37.855713 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfe81aba-0bca-45d3-96f4-fd6ee0fe7580" containerName="init" Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.855731 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfe81aba-0bca-45d3-96f4-fd6ee0fe7580" containerName="init" Dec 03 16:22:37 crc kubenswrapper[4998]: E1203 16:22:37.855789 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfe81aba-0bca-45d3-96f4-fd6ee0fe7580" containerName="dnsmasq-dns" Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.855798 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfe81aba-0bca-45d3-96f4-fd6ee0fe7580" containerName="dnsmasq-dns" Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.857273 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfe81aba-0bca-45d3-96f4-fd6ee0fe7580" containerName="dnsmasq-dns" Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.886155 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.886506 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.888887 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.888911 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-flx6d" Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.889031 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 03 16:22:37 crc kubenswrapper[4998]: I1203 16:22:37.889341 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 03 16:22:38 crc kubenswrapper[4998]: I1203 16:22:38.008035 4998 scope.go:117] "RemoveContainer" containerID="a51dec14a58fa537bd3e31f8c8880f0ddf874a18fa4fcb92cda4d91716e7b04d" Dec 03 16:22:38 crc kubenswrapper[4998]: E1203 16:22:38.014928 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a51dec14a58fa537bd3e31f8c8880f0ddf874a18fa4fcb92cda4d91716e7b04d\": container with ID starting with a51dec14a58fa537bd3e31f8c8880f0ddf874a18fa4fcb92cda4d91716e7b04d not found: ID does not exist" containerID="a51dec14a58fa537bd3e31f8c8880f0ddf874a18fa4fcb92cda4d91716e7b04d" Dec 03 16:22:38 crc kubenswrapper[4998]: I1203 16:22:38.014970 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a51dec14a58fa537bd3e31f8c8880f0ddf874a18fa4fcb92cda4d91716e7b04d"} err="failed to get container status \"a51dec14a58fa537bd3e31f8c8880f0ddf874a18fa4fcb92cda4d91716e7b04d\": rpc error: code = NotFound desc = could not find container \"a51dec14a58fa537bd3e31f8c8880f0ddf874a18fa4fcb92cda4d91716e7b04d\": container with ID starting with a51dec14a58fa537bd3e31f8c8880f0ddf874a18fa4fcb92cda4d91716e7b04d not found: ID does not exist" Dec 03 16:22:38 crc kubenswrapper[4998]: I1203 16:22:38.014996 4998 scope.go:117] "RemoveContainer" containerID="b7922d628f552d552b3534838a2a83c5ec05bc127893b947c0d2d23f6ca59187" Dec 03 16:22:38 crc kubenswrapper[4998]: E1203 16:22:38.023999 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7922d628f552d552b3534838a2a83c5ec05bc127893b947c0d2d23f6ca59187\": container with ID starting with b7922d628f552d552b3534838a2a83c5ec05bc127893b947c0d2d23f6ca59187 not found: ID does not exist" containerID="b7922d628f552d552b3534838a2a83c5ec05bc127893b947c0d2d23f6ca59187" Dec 03 16:22:38 crc kubenswrapper[4998]: I1203 16:22:38.024043 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7922d628f552d552b3534838a2a83c5ec05bc127893b947c0d2d23f6ca59187"} err="failed to get container status \"b7922d628f552d552b3534838a2a83c5ec05bc127893b947c0d2d23f6ca59187\": rpc error: code = NotFound desc = could not find container \"b7922d628f552d552b3534838a2a83c5ec05bc127893b947c0d2d23f6ca59187\": container with ID starting with b7922d628f552d552b3534838a2a83c5ec05bc127893b947c0d2d23f6ca59187 not found: ID does not exist" Dec 03 16:22:38 crc kubenswrapper[4998]: I1203 16:22:38.037957 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/d01e9a78-9e2c-44c5-9602-4c8860123f88-lock\") pod \"swift-storage-0\" (UID: \"d01e9a78-9e2c-44c5-9602-4c8860123f88\") " pod="openstack/swift-storage-0" Dec 03 16:22:38 crc kubenswrapper[4998]: I1203 16:22:38.038059 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d01e9a78-9e2c-44c5-9602-4c8860123f88-etc-swift\") pod \"swift-storage-0\" (UID: \"d01e9a78-9e2c-44c5-9602-4c8860123f88\") " pod="openstack/swift-storage-0" Dec 03 16:22:38 crc kubenswrapper[4998]: I1203 16:22:38.038101 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"d01e9a78-9e2c-44c5-9602-4c8860123f88\") " pod="openstack/swift-storage-0" Dec 03 16:22:38 crc kubenswrapper[4998]: I1203 16:22:38.038129 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/d01e9a78-9e2c-44c5-9602-4c8860123f88-cache\") pod \"swift-storage-0\" (UID: \"d01e9a78-9e2c-44c5-9602-4c8860123f88\") " pod="openstack/swift-storage-0" Dec 03 16:22:38 crc kubenswrapper[4998]: I1203 16:22:38.038145 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c897m\" (UniqueName: \"kubernetes.io/projected/d01e9a78-9e2c-44c5-9602-4c8860123f88-kube-api-access-c897m\") pod \"swift-storage-0\" (UID: \"d01e9a78-9e2c-44c5-9602-4c8860123f88\") " pod="openstack/swift-storage-0" Dec 03 16:22:38 crc kubenswrapper[4998]: I1203 16:22:38.140032 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/d01e9a78-9e2c-44c5-9602-4c8860123f88-cache\") pod \"swift-storage-0\" (UID: \"d01e9a78-9e2c-44c5-9602-4c8860123f88\") " pod="openstack/swift-storage-0" Dec 03 16:22:38 crc kubenswrapper[4998]: I1203 16:22:38.140079 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c897m\" (UniqueName: \"kubernetes.io/projected/d01e9a78-9e2c-44c5-9602-4c8860123f88-kube-api-access-c897m\") pod \"swift-storage-0\" (UID: \"d01e9a78-9e2c-44c5-9602-4c8860123f88\") " pod="openstack/swift-storage-0" Dec 03 16:22:38 crc kubenswrapper[4998]: I1203 16:22:38.140117 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/d01e9a78-9e2c-44c5-9602-4c8860123f88-lock\") pod \"swift-storage-0\" (UID: \"d01e9a78-9e2c-44c5-9602-4c8860123f88\") " pod="openstack/swift-storage-0" Dec 03 16:22:38 crc kubenswrapper[4998]: I1203 16:22:38.140194 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d01e9a78-9e2c-44c5-9602-4c8860123f88-etc-swift\") pod \"swift-storage-0\" (UID: \"d01e9a78-9e2c-44c5-9602-4c8860123f88\") " pod="openstack/swift-storage-0" Dec 03 16:22:38 crc kubenswrapper[4998]: I1203 16:22:38.140231 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"d01e9a78-9e2c-44c5-9602-4c8860123f88\") " pod="openstack/swift-storage-0" Dec 03 16:22:38 crc kubenswrapper[4998]: E1203 16:22:38.140520 4998 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 16:22:38 crc kubenswrapper[4998]: E1203 16:22:38.140541 4998 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 16:22:38 crc kubenswrapper[4998]: E1203 16:22:38.140578 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d01e9a78-9e2c-44c5-9602-4c8860123f88-etc-swift podName:d01e9a78-9e2c-44c5-9602-4c8860123f88 nodeName:}" failed. No retries permitted until 2025-12-03 16:22:38.640564253 +0000 UTC m=+1137.252264476 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d01e9a78-9e2c-44c5-9602-4c8860123f88-etc-swift") pod "swift-storage-0" (UID: "d01e9a78-9e2c-44c5-9602-4c8860123f88") : configmap "swift-ring-files" not found Dec 03 16:22:38 crc kubenswrapper[4998]: I1203 16:22:38.140614 4998 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"d01e9a78-9e2c-44c5-9602-4c8860123f88\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/swift-storage-0" Dec 03 16:22:38 crc kubenswrapper[4998]: I1203 16:22:38.140670 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/d01e9a78-9e2c-44c5-9602-4c8860123f88-cache\") pod \"swift-storage-0\" (UID: \"d01e9a78-9e2c-44c5-9602-4c8860123f88\") " pod="openstack/swift-storage-0" Dec 03 16:22:38 crc kubenswrapper[4998]: I1203 16:22:38.140972 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/d01e9a78-9e2c-44c5-9602-4c8860123f88-lock\") pod \"swift-storage-0\" (UID: \"d01e9a78-9e2c-44c5-9602-4c8860123f88\") " pod="openstack/swift-storage-0" Dec 03 16:22:38 crc kubenswrapper[4998]: I1203 16:22:38.159200 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c897m\" (UniqueName: \"kubernetes.io/projected/d01e9a78-9e2c-44c5-9602-4c8860123f88-kube-api-access-c897m\") pod \"swift-storage-0\" (UID: \"d01e9a78-9e2c-44c5-9602-4c8860123f88\") " pod="openstack/swift-storage-0" Dec 03 16:22:38 crc kubenswrapper[4998]: I1203 16:22:38.166201 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"d01e9a78-9e2c-44c5-9602-4c8860123f88\") " pod="openstack/swift-storage-0" Dec 03 16:22:38 crc kubenswrapper[4998]: E1203 16:22:38.646466 4998 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode266f963_33c9_48dc_bb10_70bd5c700d69.slice/crio-241d5701046d2695d1096b4ebf57a93394e22c261fa3f23621e2e3b2ec074ffc.scope\": RecentStats: unable to find data in memory cache]" Dec 03 16:22:38 crc kubenswrapper[4998]: I1203 16:22:38.647973 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d01e9a78-9e2c-44c5-9602-4c8860123f88-etc-swift\") pod \"swift-storage-0\" (UID: \"d01e9a78-9e2c-44c5-9602-4c8860123f88\") " pod="openstack/swift-storage-0" Dec 03 16:22:38 crc kubenswrapper[4998]: E1203 16:22:38.648173 4998 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 16:22:38 crc kubenswrapper[4998]: E1203 16:22:38.648204 4998 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 16:22:38 crc kubenswrapper[4998]: E1203 16:22:38.648258 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d01e9a78-9e2c-44c5-9602-4c8860123f88-etc-swift podName:d01e9a78-9e2c-44c5-9602-4c8860123f88 nodeName:}" failed. No retries permitted until 2025-12-03 16:22:39.648235943 +0000 UTC m=+1138.259936166 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d01e9a78-9e2c-44c5-9602-4c8860123f88-etc-swift") pod "swift-storage-0" (UID: "d01e9a78-9e2c-44c5-9602-4c8860123f88") : configmap "swift-ring-files" not found Dec 03 16:22:38 crc kubenswrapper[4998]: I1203 16:22:38.722143 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" event={"ID":"e266f963-33c9-48dc-bb10-70bd5c700d69","Type":"ContainerStarted","Data":"241d5701046d2695d1096b4ebf57a93394e22c261fa3f23621e2e3b2ec074ffc"} Dec 03 16:22:38 crc kubenswrapper[4998]: I1203 16:22:38.722198 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" event={"ID":"e266f963-33c9-48dc-bb10-70bd5c700d69","Type":"ContainerStarted","Data":"7e6b8fcb2c28749e83376531d1ca7c381481f59b2a757cb53d68873facac9224"} Dec 03 16:22:39 crc kubenswrapper[4998]: I1203 16:22:39.149631 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 03 16:22:39 crc kubenswrapper[4998]: I1203 16:22:39.303711 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 03 16:22:39 crc kubenswrapper[4998]: I1203 16:22:39.304921 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b6fcbcc87-jjnql" Dec 03 16:22:39 crc kubenswrapper[4998]: I1203 16:22:39.675484 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d01e9a78-9e2c-44c5-9602-4c8860123f88-etc-swift\") pod \"swift-storage-0\" (UID: \"d01e9a78-9e2c-44c5-9602-4c8860123f88\") " pod="openstack/swift-storage-0" Dec 03 16:22:39 crc kubenswrapper[4998]: E1203 16:22:39.675723 4998 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 16:22:39 crc kubenswrapper[4998]: E1203 16:22:39.675777 4998 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 16:22:39 crc kubenswrapper[4998]: E1203 16:22:39.675835 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d01e9a78-9e2c-44c5-9602-4c8860123f88-etc-swift podName:d01e9a78-9e2c-44c5-9602-4c8860123f88 nodeName:}" failed. No retries permitted until 2025-12-03 16:22:41.67581676 +0000 UTC m=+1140.287516983 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d01e9a78-9e2c-44c5-9602-4c8860123f88-etc-swift") pod "swift-storage-0" (UID: "d01e9a78-9e2c-44c5-9602-4c8860123f88") : configmap "swift-ring-files" not found Dec 03 16:22:39 crc kubenswrapper[4998]: I1203 16:22:39.687497 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfe81aba-0bca-45d3-96f4-fd6ee0fe7580" path="/var/lib/kubelet/pods/dfe81aba-0bca-45d3-96f4-fd6ee0fe7580/volumes" Dec 03 16:22:39 crc kubenswrapper[4998]: I1203 16:22:39.733264 4998 generic.go:334] "Generic (PLEG): container finished" podID="e266f963-33c9-48dc-bb10-70bd5c700d69" containerID="241d5701046d2695d1096b4ebf57a93394e22c261fa3f23621e2e3b2ec074ffc" exitCode=0 Dec 03 16:22:39 crc kubenswrapper[4998]: I1203 16:22:39.733360 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" event={"ID":"e266f963-33c9-48dc-bb10-70bd5c700d69","Type":"ContainerDied","Data":"241d5701046d2695d1096b4ebf57a93394e22c261fa3f23621e2e3b2ec074ffc"} Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.707783 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d01e9a78-9e2c-44c5-9602-4c8860123f88-etc-swift\") pod \"swift-storage-0\" (UID: \"d01e9a78-9e2c-44c5-9602-4c8860123f88\") " pod="openstack/swift-storage-0" Dec 03 16:22:41 crc kubenswrapper[4998]: E1203 16:22:41.708444 4998 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 16:22:41 crc kubenswrapper[4998]: E1203 16:22:41.708466 4998 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 16:22:41 crc kubenswrapper[4998]: E1203 16:22:41.708516 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d01e9a78-9e2c-44c5-9602-4c8860123f88-etc-swift podName:d01e9a78-9e2c-44c5-9602-4c8860123f88 nodeName:}" failed. No retries permitted until 2025-12-03 16:22:45.70850137 +0000 UTC m=+1144.320201593 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d01e9a78-9e2c-44c5-9602-4c8860123f88-etc-swift") pod "swift-storage-0" (UID: "d01e9a78-9e2c-44c5-9602-4c8860123f88") : configmap "swift-ring-files" not found Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.766326 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-xn4z8"] Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.767376 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-xn4z8" Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.770993 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.771370 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.773072 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.785680 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-xn4z8"] Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.809927 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-swiftconf\") pod \"swift-ring-rebalance-xn4z8\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " pod="openstack/swift-ring-rebalance-xn4z8" Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.809996 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-etc-swift\") pod \"swift-ring-rebalance-xn4z8\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " pod="openstack/swift-ring-rebalance-xn4z8" Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.810061 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-scripts\") pod \"swift-ring-rebalance-xn4z8\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " pod="openstack/swift-ring-rebalance-xn4z8" Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.810128 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-ring-data-devices\") pod \"swift-ring-rebalance-xn4z8\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " pod="openstack/swift-ring-rebalance-xn4z8" Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.810163 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7vj2\" (UniqueName: \"kubernetes.io/projected/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-kube-api-access-q7vj2\") pod \"swift-ring-rebalance-xn4z8\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " pod="openstack/swift-ring-rebalance-xn4z8" Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.810199 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-combined-ca-bundle\") pod \"swift-ring-rebalance-xn4z8\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " pod="openstack/swift-ring-rebalance-xn4z8" Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.810224 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-dispersionconf\") pod \"swift-ring-rebalance-xn4z8\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " pod="openstack/swift-ring-rebalance-xn4z8" Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.912140 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-ring-data-devices\") pod \"swift-ring-rebalance-xn4z8\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " pod="openstack/swift-ring-rebalance-xn4z8" Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.912193 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7vj2\" (UniqueName: \"kubernetes.io/projected/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-kube-api-access-q7vj2\") pod \"swift-ring-rebalance-xn4z8\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " pod="openstack/swift-ring-rebalance-xn4z8" Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.912228 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-combined-ca-bundle\") pod \"swift-ring-rebalance-xn4z8\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " pod="openstack/swift-ring-rebalance-xn4z8" Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.912253 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-dispersionconf\") pod \"swift-ring-rebalance-xn4z8\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " pod="openstack/swift-ring-rebalance-xn4z8" Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.912323 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-swiftconf\") pod \"swift-ring-rebalance-xn4z8\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " pod="openstack/swift-ring-rebalance-xn4z8" Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.912354 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-etc-swift\") pod \"swift-ring-rebalance-xn4z8\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " pod="openstack/swift-ring-rebalance-xn4z8" Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.912396 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-scripts\") pod \"swift-ring-rebalance-xn4z8\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " pod="openstack/swift-ring-rebalance-xn4z8" Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.913711 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-etc-swift\") pod \"swift-ring-rebalance-xn4z8\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " pod="openstack/swift-ring-rebalance-xn4z8" Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.913846 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-ring-data-devices\") pod \"swift-ring-rebalance-xn4z8\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " pod="openstack/swift-ring-rebalance-xn4z8" Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.913975 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-scripts\") pod \"swift-ring-rebalance-xn4z8\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " pod="openstack/swift-ring-rebalance-xn4z8" Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.919256 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-dispersionconf\") pod \"swift-ring-rebalance-xn4z8\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " pod="openstack/swift-ring-rebalance-xn4z8" Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.919569 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-combined-ca-bundle\") pod \"swift-ring-rebalance-xn4z8\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " pod="openstack/swift-ring-rebalance-xn4z8" Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.920121 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-swiftconf\") pod \"swift-ring-rebalance-xn4z8\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " pod="openstack/swift-ring-rebalance-xn4z8" Dec 03 16:22:41 crc kubenswrapper[4998]: I1203 16:22:41.928362 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7vj2\" (UniqueName: \"kubernetes.io/projected/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-kube-api-access-q7vj2\") pod \"swift-ring-rebalance-xn4z8\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " pod="openstack/swift-ring-rebalance-xn4z8" Dec 03 16:22:42 crc kubenswrapper[4998]: I1203 16:22:42.088731 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-flx6d" Dec 03 16:22:42 crc kubenswrapper[4998]: I1203 16:22:42.095556 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-xn4z8" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.099395 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-daeb-account-create-update-vsrjb"] Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.101850 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-daeb-account-create-update-vsrjb" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.104376 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.143663 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-daeb-account-create-update-vsrjb"] Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.152743 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqfxt\" (UniqueName: \"kubernetes.io/projected/b542f7b1-ca06-4b56-b45c-dccde2fca019-kube-api-access-nqfxt\") pod \"keystone-daeb-account-create-update-vsrjb\" (UID: \"b542f7b1-ca06-4b56-b45c-dccde2fca019\") " pod="openstack/keystone-daeb-account-create-update-vsrjb" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.153291 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b542f7b1-ca06-4b56-b45c-dccde2fca019-operator-scripts\") pod \"keystone-daeb-account-create-update-vsrjb\" (UID: \"b542f7b1-ca06-4b56-b45c-dccde2fca019\") " pod="openstack/keystone-daeb-account-create-update-vsrjb" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.153959 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-gwldb"] Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.156607 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gwldb" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.184635 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-gwldb"] Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.255602 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b542f7b1-ca06-4b56-b45c-dccde2fca019-operator-scripts\") pod \"keystone-daeb-account-create-update-vsrjb\" (UID: \"b542f7b1-ca06-4b56-b45c-dccde2fca019\") " pod="openstack/keystone-daeb-account-create-update-vsrjb" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.255724 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ce04863-7333-4484-9507-93143d276e47-operator-scripts\") pod \"keystone-db-create-gwldb\" (UID: \"0ce04863-7333-4484-9507-93143d276e47\") " pod="openstack/keystone-db-create-gwldb" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.255794 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqfxt\" (UniqueName: \"kubernetes.io/projected/b542f7b1-ca06-4b56-b45c-dccde2fca019-kube-api-access-nqfxt\") pod \"keystone-daeb-account-create-update-vsrjb\" (UID: \"b542f7b1-ca06-4b56-b45c-dccde2fca019\") " pod="openstack/keystone-daeb-account-create-update-vsrjb" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.255825 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lldh\" (UniqueName: \"kubernetes.io/projected/0ce04863-7333-4484-9507-93143d276e47-kube-api-access-2lldh\") pod \"keystone-db-create-gwldb\" (UID: \"0ce04863-7333-4484-9507-93143d276e47\") " pod="openstack/keystone-db-create-gwldb" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.256590 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b542f7b1-ca06-4b56-b45c-dccde2fca019-operator-scripts\") pod \"keystone-daeb-account-create-update-vsrjb\" (UID: \"b542f7b1-ca06-4b56-b45c-dccde2fca019\") " pod="openstack/keystone-daeb-account-create-update-vsrjb" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.277147 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqfxt\" (UniqueName: \"kubernetes.io/projected/b542f7b1-ca06-4b56-b45c-dccde2fca019-kube-api-access-nqfxt\") pod \"keystone-daeb-account-create-update-vsrjb\" (UID: \"b542f7b1-ca06-4b56-b45c-dccde2fca019\") " pod="openstack/keystone-daeb-account-create-update-vsrjb" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.358113 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ce04863-7333-4484-9507-93143d276e47-operator-scripts\") pod \"keystone-db-create-gwldb\" (UID: \"0ce04863-7333-4484-9507-93143d276e47\") " pod="openstack/keystone-db-create-gwldb" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.358174 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lldh\" (UniqueName: \"kubernetes.io/projected/0ce04863-7333-4484-9507-93143d276e47-kube-api-access-2lldh\") pod \"keystone-db-create-gwldb\" (UID: \"0ce04863-7333-4484-9507-93143d276e47\") " pod="openstack/keystone-db-create-gwldb" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.359038 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ce04863-7333-4484-9507-93143d276e47-operator-scripts\") pod \"keystone-db-create-gwldb\" (UID: \"0ce04863-7333-4484-9507-93143d276e47\") " pod="openstack/keystone-db-create-gwldb" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.377898 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lldh\" (UniqueName: \"kubernetes.io/projected/0ce04863-7333-4484-9507-93143d276e47-kube-api-access-2lldh\") pod \"keystone-db-create-gwldb\" (UID: \"0ce04863-7333-4484-9507-93143d276e47\") " pod="openstack/keystone-db-create-gwldb" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.380382 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-p4448"] Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.381586 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-p4448" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.393487 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-p4448"] Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.457874 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-daeb-account-create-update-vsrjb" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.459877 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zjpz\" (UniqueName: \"kubernetes.io/projected/3739b197-1e59-48fa-a7f7-9e73b95a382a-kube-api-access-4zjpz\") pod \"placement-db-create-p4448\" (UID: \"3739b197-1e59-48fa-a7f7-9e73b95a382a\") " pod="openstack/placement-db-create-p4448" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.460158 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3739b197-1e59-48fa-a7f7-9e73b95a382a-operator-scripts\") pod \"placement-db-create-p4448\" (UID: \"3739b197-1e59-48fa-a7f7-9e73b95a382a\") " pod="openstack/placement-db-create-p4448" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.491697 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gwldb" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.510719 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-fffa-account-create-update-bh4qr"] Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.511921 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-fffa-account-create-update-bh4qr" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.517583 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.520121 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-fffa-account-create-update-bh4qr"] Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.561791 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3739b197-1e59-48fa-a7f7-9e73b95a382a-operator-scripts\") pod \"placement-db-create-p4448\" (UID: \"3739b197-1e59-48fa-a7f7-9e73b95a382a\") " pod="openstack/placement-db-create-p4448" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.561852 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s27v9\" (UniqueName: \"kubernetes.io/projected/72129c14-7e4f-4e0a-9053-7d2178a835ae-kube-api-access-s27v9\") pod \"placement-fffa-account-create-update-bh4qr\" (UID: \"72129c14-7e4f-4e0a-9053-7d2178a835ae\") " pod="openstack/placement-fffa-account-create-update-bh4qr" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.561875 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72129c14-7e4f-4e0a-9053-7d2178a835ae-operator-scripts\") pod \"placement-fffa-account-create-update-bh4qr\" (UID: \"72129c14-7e4f-4e0a-9053-7d2178a835ae\") " pod="openstack/placement-fffa-account-create-update-bh4qr" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.561976 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zjpz\" (UniqueName: \"kubernetes.io/projected/3739b197-1e59-48fa-a7f7-9e73b95a382a-kube-api-access-4zjpz\") pod \"placement-db-create-p4448\" (UID: \"3739b197-1e59-48fa-a7f7-9e73b95a382a\") " pod="openstack/placement-db-create-p4448" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.562915 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3739b197-1e59-48fa-a7f7-9e73b95a382a-operator-scripts\") pod \"placement-db-create-p4448\" (UID: \"3739b197-1e59-48fa-a7f7-9e73b95a382a\") " pod="openstack/placement-db-create-p4448" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.578040 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zjpz\" (UniqueName: \"kubernetes.io/projected/3739b197-1e59-48fa-a7f7-9e73b95a382a-kube-api-access-4zjpz\") pod \"placement-db-create-p4448\" (UID: \"3739b197-1e59-48fa-a7f7-9e73b95a382a\") " pod="openstack/placement-db-create-p4448" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.663059 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s27v9\" (UniqueName: \"kubernetes.io/projected/72129c14-7e4f-4e0a-9053-7d2178a835ae-kube-api-access-s27v9\") pod \"placement-fffa-account-create-update-bh4qr\" (UID: \"72129c14-7e4f-4e0a-9053-7d2178a835ae\") " pod="openstack/placement-fffa-account-create-update-bh4qr" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.663107 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72129c14-7e4f-4e0a-9053-7d2178a835ae-operator-scripts\") pod \"placement-fffa-account-create-update-bh4qr\" (UID: \"72129c14-7e4f-4e0a-9053-7d2178a835ae\") " pod="openstack/placement-fffa-account-create-update-bh4qr" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.663799 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72129c14-7e4f-4e0a-9053-7d2178a835ae-operator-scripts\") pod \"placement-fffa-account-create-update-bh4qr\" (UID: \"72129c14-7e4f-4e0a-9053-7d2178a835ae\") " pod="openstack/placement-fffa-account-create-update-bh4qr" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.683827 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s27v9\" (UniqueName: \"kubernetes.io/projected/72129c14-7e4f-4e0a-9053-7d2178a835ae-kube-api-access-s27v9\") pod \"placement-fffa-account-create-update-bh4qr\" (UID: \"72129c14-7e4f-4e0a-9053-7d2178a835ae\") " pod="openstack/placement-fffa-account-create-update-bh4qr" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.743087 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-p4448" Dec 03 16:22:44 crc kubenswrapper[4998]: I1203 16:22:44.833811 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-fffa-account-create-update-bh4qr" Dec 03 16:22:45 crc kubenswrapper[4998]: I1203 16:22:45.784088 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d01e9a78-9e2c-44c5-9602-4c8860123f88-etc-swift\") pod \"swift-storage-0\" (UID: \"d01e9a78-9e2c-44c5-9602-4c8860123f88\") " pod="openstack/swift-storage-0" Dec 03 16:22:45 crc kubenswrapper[4998]: E1203 16:22:45.784343 4998 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 16:22:45 crc kubenswrapper[4998]: E1203 16:22:45.784371 4998 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 16:22:45 crc kubenswrapper[4998]: E1203 16:22:45.784435 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d01e9a78-9e2c-44c5-9602-4c8860123f88-etc-swift podName:d01e9a78-9e2c-44c5-9602-4c8860123f88 nodeName:}" failed. No retries permitted until 2025-12-03 16:22:53.784412805 +0000 UTC m=+1152.396113108 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d01e9a78-9e2c-44c5-9602-4c8860123f88-etc-swift") pod "swift-storage-0" (UID: "d01e9a78-9e2c-44c5-9602-4c8860123f88") : configmap "swift-ring-files" not found Dec 03 16:22:46 crc kubenswrapper[4998]: I1203 16:22:46.698175 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-db-create-r4skm"] Dec 03 16:22:46 crc kubenswrapper[4998]: I1203 16:22:46.700260 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-r4skm" Dec 03 16:22:46 crc kubenswrapper[4998]: I1203 16:22:46.711595 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-create-r4skm"] Dec 03 16:22:46 crc kubenswrapper[4998]: I1203 16:22:46.793236 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-21ed-account-create-update-qnv6c"] Dec 03 16:22:46 crc kubenswrapper[4998]: I1203 16:22:46.794382 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-21ed-account-create-update-qnv6c" Dec 03 16:22:46 crc kubenswrapper[4998]: I1203 16:22:46.796696 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-db-secret" Dec 03 16:22:46 crc kubenswrapper[4998]: I1203 16:22:46.806575 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48d7v\" (UniqueName: \"kubernetes.io/projected/d71c5697-f3d2-4c29-a722-49823fa55d6d-kube-api-access-48d7v\") pod \"watcher-db-create-r4skm\" (UID: \"d71c5697-f3d2-4c29-a722-49823fa55d6d\") " pod="openstack/watcher-db-create-r4skm" Dec 03 16:22:46 crc kubenswrapper[4998]: I1203 16:22:46.806707 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d71c5697-f3d2-4c29-a722-49823fa55d6d-operator-scripts\") pod \"watcher-db-create-r4skm\" (UID: \"d71c5697-f3d2-4c29-a722-49823fa55d6d\") " pod="openstack/watcher-db-create-r4skm" Dec 03 16:22:46 crc kubenswrapper[4998]: I1203 16:22:46.807164 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-21ed-account-create-update-qnv6c"] Dec 03 16:22:46 crc kubenswrapper[4998]: I1203 16:22:46.908840 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48d7v\" (UniqueName: \"kubernetes.io/projected/d71c5697-f3d2-4c29-a722-49823fa55d6d-kube-api-access-48d7v\") pod \"watcher-db-create-r4skm\" (UID: \"d71c5697-f3d2-4c29-a722-49823fa55d6d\") " pod="openstack/watcher-db-create-r4skm" Dec 03 16:22:46 crc kubenswrapper[4998]: I1203 16:22:46.908999 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15941ab6-753a-4eb7-8e6f-51ffae3a0227-operator-scripts\") pod \"watcher-21ed-account-create-update-qnv6c\" (UID: \"15941ab6-753a-4eb7-8e6f-51ffae3a0227\") " pod="openstack/watcher-21ed-account-create-update-qnv6c" Dec 03 16:22:46 crc kubenswrapper[4998]: I1203 16:22:46.909086 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d71c5697-f3d2-4c29-a722-49823fa55d6d-operator-scripts\") pod \"watcher-db-create-r4skm\" (UID: \"d71c5697-f3d2-4c29-a722-49823fa55d6d\") " pod="openstack/watcher-db-create-r4skm" Dec 03 16:22:46 crc kubenswrapper[4998]: I1203 16:22:46.909126 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ml7z\" (UniqueName: \"kubernetes.io/projected/15941ab6-753a-4eb7-8e6f-51ffae3a0227-kube-api-access-8ml7z\") pod \"watcher-21ed-account-create-update-qnv6c\" (UID: \"15941ab6-753a-4eb7-8e6f-51ffae3a0227\") " pod="openstack/watcher-21ed-account-create-update-qnv6c" Dec 03 16:22:46 crc kubenswrapper[4998]: I1203 16:22:46.910045 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d71c5697-f3d2-4c29-a722-49823fa55d6d-operator-scripts\") pod \"watcher-db-create-r4skm\" (UID: \"d71c5697-f3d2-4c29-a722-49823fa55d6d\") " pod="openstack/watcher-db-create-r4skm" Dec 03 16:22:46 crc kubenswrapper[4998]: I1203 16:22:46.939085 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48d7v\" (UniqueName: \"kubernetes.io/projected/d71c5697-f3d2-4c29-a722-49823fa55d6d-kube-api-access-48d7v\") pod \"watcher-db-create-r4skm\" (UID: \"d71c5697-f3d2-4c29-a722-49823fa55d6d\") " pod="openstack/watcher-db-create-r4skm" Dec 03 16:22:47 crc kubenswrapper[4998]: I1203 16:22:47.011400 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ml7z\" (UniqueName: \"kubernetes.io/projected/15941ab6-753a-4eb7-8e6f-51ffae3a0227-kube-api-access-8ml7z\") pod \"watcher-21ed-account-create-update-qnv6c\" (UID: \"15941ab6-753a-4eb7-8e6f-51ffae3a0227\") " pod="openstack/watcher-21ed-account-create-update-qnv6c" Dec 03 16:22:47 crc kubenswrapper[4998]: I1203 16:22:47.011971 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15941ab6-753a-4eb7-8e6f-51ffae3a0227-operator-scripts\") pod \"watcher-21ed-account-create-update-qnv6c\" (UID: \"15941ab6-753a-4eb7-8e6f-51ffae3a0227\") " pod="openstack/watcher-21ed-account-create-update-qnv6c" Dec 03 16:22:47 crc kubenswrapper[4998]: I1203 16:22:47.012880 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15941ab6-753a-4eb7-8e6f-51ffae3a0227-operator-scripts\") pod \"watcher-21ed-account-create-update-qnv6c\" (UID: \"15941ab6-753a-4eb7-8e6f-51ffae3a0227\") " pod="openstack/watcher-21ed-account-create-update-qnv6c" Dec 03 16:22:47 crc kubenswrapper[4998]: I1203 16:22:47.043102 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ml7z\" (UniqueName: \"kubernetes.io/projected/15941ab6-753a-4eb7-8e6f-51ffae3a0227-kube-api-access-8ml7z\") pod \"watcher-21ed-account-create-update-qnv6c\" (UID: \"15941ab6-753a-4eb7-8e6f-51ffae3a0227\") " pod="openstack/watcher-21ed-account-create-update-qnv6c" Dec 03 16:22:47 crc kubenswrapper[4998]: I1203 16:22:47.048057 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-r4skm" Dec 03 16:22:47 crc kubenswrapper[4998]: I1203 16:22:47.120980 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-21ed-account-create-update-qnv6c" Dec 03 16:22:49 crc kubenswrapper[4998]: I1203 16:22:49.347983 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 03 16:22:53 crc kubenswrapper[4998]: I1203 16:22:53.324634 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 03 16:22:53 crc kubenswrapper[4998]: I1203 16:22:53.473141 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 03 16:22:53 crc kubenswrapper[4998]: I1203 16:22:53.754650 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-p4448"] Dec 03 16:22:53 crc kubenswrapper[4998]: I1203 16:22:53.850979 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" event={"ID":"e266f963-33c9-48dc-bb10-70bd5c700d69","Type":"ContainerStarted","Data":"8a02ff22240d153f314a59c60dd4559c01cd622cfce7f13356e6514f03fde803"} Dec 03 16:22:53 crc kubenswrapper[4998]: I1203 16:22:53.852236 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" Dec 03 16:22:53 crc kubenswrapper[4998]: I1203 16:22:53.854041 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d01e9a78-9e2c-44c5-9602-4c8860123f88-etc-swift\") pod \"swift-storage-0\" (UID: \"d01e9a78-9e2c-44c5-9602-4c8860123f88\") " pod="openstack/swift-storage-0" Dec 03 16:22:53 crc kubenswrapper[4998]: E1203 16:22:53.854281 4998 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 16:22:53 crc kubenswrapper[4998]: E1203 16:22:53.854297 4998 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 16:22:53 crc kubenswrapper[4998]: E1203 16:22:53.854333 4998 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d01e9a78-9e2c-44c5-9602-4c8860123f88-etc-swift podName:d01e9a78-9e2c-44c5-9602-4c8860123f88 nodeName:}" failed. No retries permitted until 2025-12-03 16:23:09.854320564 +0000 UTC m=+1168.466020777 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d01e9a78-9e2c-44c5-9602-4c8860123f88-etc-swift") pod "swift-storage-0" (UID: "d01e9a78-9e2c-44c5-9602-4c8860123f88") : configmap "swift-ring-files" not found Dec 03 16:22:53 crc kubenswrapper[4998]: I1203 16:22:53.855278 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-p4448" event={"ID":"3739b197-1e59-48fa-a7f7-9e73b95a382a","Type":"ContainerStarted","Data":"81b6bcecbb278d74851fe2744b5be1b0c56a1106d4418c169feb33670878badc"} Dec 03 16:22:53 crc kubenswrapper[4998]: I1203 16:22:53.857469 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92d2d9f9-b0c3-40f8-981a-41ec47f030e1","Type":"ContainerStarted","Data":"39b4425a9a234b6550ab87b9df86280de262b2abf3cd4770ea0623e4f4b32977"} Dec 03 16:22:53 crc kubenswrapper[4998]: I1203 16:22:53.876509 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" podStartSLOduration=17.876492363 podStartE2EDuration="17.876492363s" podCreationTimestamp="2025-12-03 16:22:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:22:53.871059716 +0000 UTC m=+1152.482759939" watchObservedRunningTime="2025-12-03 16:22:53.876492363 +0000 UTC m=+1152.488192586" Dec 03 16:22:53 crc kubenswrapper[4998]: W1203 16:22:53.994407 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ce04863_7333_4484_9507_93143d276e47.slice/crio-566908bc977b146ab72ce9e307b7202dc74e0053df3769c5a7dbd6f1b3329fb7 WatchSource:0}: Error finding container 566908bc977b146ab72ce9e307b7202dc74e0053df3769c5a7dbd6f1b3329fb7: Status 404 returned error can't find the container with id 566908bc977b146ab72ce9e307b7202dc74e0053df3769c5a7dbd6f1b3329fb7 Dec 03 16:22:54 crc kubenswrapper[4998]: I1203 16:22:54.003371 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-gwldb"] Dec 03 16:22:54 crc kubenswrapper[4998]: I1203 16:22:54.023960 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-fffa-account-create-update-bh4qr"] Dec 03 16:22:54 crc kubenswrapper[4998]: W1203 16:22:54.026383 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb542f7b1_ca06_4b56_b45c_dccde2fca019.slice/crio-501ae843f2c6c45ed3119d58a1e93d5c2428b59ad584d62b13128c74cd647922 WatchSource:0}: Error finding container 501ae843f2c6c45ed3119d58a1e93d5c2428b59ad584d62b13128c74cd647922: Status 404 returned error can't find the container with id 501ae843f2c6c45ed3119d58a1e93d5c2428b59ad584d62b13128c74cd647922 Dec 03 16:22:54 crc kubenswrapper[4998]: W1203 16:22:54.029236 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72129c14_7e4f_4e0a_9053_7d2178a835ae.slice/crio-70d9067c3c9c8379cb947632219ffc0bcd05917396ad73334e0da048bb2fd8f4 WatchSource:0}: Error finding container 70d9067c3c9c8379cb947632219ffc0bcd05917396ad73334e0da048bb2fd8f4: Status 404 returned error can't find the container with id 70d9067c3c9c8379cb947632219ffc0bcd05917396ad73334e0da048bb2fd8f4 Dec 03 16:22:54 crc kubenswrapper[4998]: I1203 16:22:54.029726 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-daeb-account-create-update-vsrjb"] Dec 03 16:22:54 crc kubenswrapper[4998]: I1203 16:22:54.141476 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-21ed-account-create-update-qnv6c"] Dec 03 16:22:54 crc kubenswrapper[4998]: W1203 16:22:54.145878 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15941ab6_753a_4eb7_8e6f_51ffae3a0227.slice/crio-ee3391e8727875e73a5391f7bd33ffa15a1459a3bcc003cca76a9d40630a9a4e WatchSource:0}: Error finding container ee3391e8727875e73a5391f7bd33ffa15a1459a3bcc003cca76a9d40630a9a4e: Status 404 returned error can't find the container with id ee3391e8727875e73a5391f7bd33ffa15a1459a3bcc003cca76a9d40630a9a4e Dec 03 16:22:54 crc kubenswrapper[4998]: I1203 16:22:54.147587 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-create-r4skm"] Dec 03 16:22:54 crc kubenswrapper[4998]: I1203 16:22:54.155595 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-xn4z8"] Dec 03 16:22:54 crc kubenswrapper[4998]: W1203 16:22:54.156600 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78d5a8fb_51f1_464f_ae8a_f5ecd2be7a0c.slice/crio-ba13d662cd465a86228e8355c796632f291f54b7fcfec3c5a6baa7bad8929396 WatchSource:0}: Error finding container ba13d662cd465a86228e8355c796632f291f54b7fcfec3c5a6baa7bad8929396: Status 404 returned error can't find the container with id ba13d662cd465a86228e8355c796632f291f54b7fcfec3c5a6baa7bad8929396 Dec 03 16:22:54 crc kubenswrapper[4998]: I1203 16:22:54.874365 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-xn4z8" event={"ID":"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c","Type":"ContainerStarted","Data":"ba13d662cd465a86228e8355c796632f291f54b7fcfec3c5a6baa7bad8929396"} Dec 03 16:22:54 crc kubenswrapper[4998]: I1203 16:22:54.876163 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-fffa-account-create-update-bh4qr" event={"ID":"72129c14-7e4f-4e0a-9053-7d2178a835ae","Type":"ContainerStarted","Data":"70d9067c3c9c8379cb947632219ffc0bcd05917396ad73334e0da048bb2fd8f4"} Dec 03 16:22:54 crc kubenswrapper[4998]: I1203 16:22:54.878566 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-p4448" event={"ID":"3739b197-1e59-48fa-a7f7-9e73b95a382a","Type":"ContainerStarted","Data":"88950ec72a9b62516073acb253b694726d83368ce8222ef79a223454a9bba802"} Dec 03 16:22:54 crc kubenswrapper[4998]: I1203 16:22:54.880917 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gwldb" event={"ID":"0ce04863-7333-4484-9507-93143d276e47","Type":"ContainerStarted","Data":"566908bc977b146ab72ce9e307b7202dc74e0053df3769c5a7dbd6f1b3329fb7"} Dec 03 16:22:54 crc kubenswrapper[4998]: I1203 16:22:54.882797 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-daeb-account-create-update-vsrjb" event={"ID":"b542f7b1-ca06-4b56-b45c-dccde2fca019","Type":"ContainerStarted","Data":"501ae843f2c6c45ed3119d58a1e93d5c2428b59ad584d62b13128c74cd647922"} Dec 03 16:22:54 crc kubenswrapper[4998]: I1203 16:22:54.885283 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-21ed-account-create-update-qnv6c" event={"ID":"15941ab6-753a-4eb7-8e6f-51ffae3a0227","Type":"ContainerStarted","Data":"ee3391e8727875e73a5391f7bd33ffa15a1459a3bcc003cca76a9d40630a9a4e"} Dec 03 16:22:54 crc kubenswrapper[4998]: I1203 16:22:54.887546 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-r4skm" event={"ID":"d71c5697-f3d2-4c29-a722-49823fa55d6d","Type":"ContainerStarted","Data":"8a3af9fd80215e1f279d7af6e8b800d5b85aa43c98e32d07d842ef59cd811258"} Dec 03 16:22:54 crc kubenswrapper[4998]: I1203 16:22:54.918683 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-p4448" podStartSLOduration=10.918658568 podStartE2EDuration="10.918658568s" podCreationTimestamp="2025-12-03 16:22:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:22:54.905620209 +0000 UTC m=+1153.517320462" watchObservedRunningTime="2025-12-03 16:22:54.918658568 +0000 UTC m=+1153.530358831" Dec 03 16:22:55 crc kubenswrapper[4998]: I1203 16:22:55.049310 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-7sxbw" podUID="95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f" containerName="ovn-controller" probeResult="failure" output=< Dec 03 16:22:55 crc kubenswrapper[4998]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 16:22:55 crc kubenswrapper[4998]: > Dec 03 16:22:55 crc kubenswrapper[4998]: I1203 16:22:55.897260 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-21ed-account-create-update-qnv6c" event={"ID":"15941ab6-753a-4eb7-8e6f-51ffae3a0227","Type":"ContainerStarted","Data":"75385576e42fd8cb783f5ce0d82d0c0520f8d6a3e21af41b30c90d912cd5f3b0"} Dec 03 16:22:55 crc kubenswrapper[4998]: I1203 16:22:55.898944 4998 generic.go:334] "Generic (PLEG): container finished" podID="b542f7b1-ca06-4b56-b45c-dccde2fca019" containerID="fedfa179097767efc0d2374a3a66b294f4cc1edfacd2dd7f9eb72f4f28be735a" exitCode=0 Dec 03 16:22:55 crc kubenswrapper[4998]: I1203 16:22:55.899009 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-daeb-account-create-update-vsrjb" event={"ID":"b542f7b1-ca06-4b56-b45c-dccde2fca019","Type":"ContainerDied","Data":"fedfa179097767efc0d2374a3a66b294f4cc1edfacd2dd7f9eb72f4f28be735a"} Dec 03 16:22:55 crc kubenswrapper[4998]: I1203 16:22:55.901348 4998 generic.go:334] "Generic (PLEG): container finished" podID="d71c5697-f3d2-4c29-a722-49823fa55d6d" containerID="32c9f837e40269fa018e70f76d3c57e75cf93e48838768e46b94217965185dbb" exitCode=0 Dec 03 16:22:55 crc kubenswrapper[4998]: I1203 16:22:55.901406 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-r4skm" event={"ID":"d71c5697-f3d2-4c29-a722-49823fa55d6d","Type":"ContainerDied","Data":"32c9f837e40269fa018e70f76d3c57e75cf93e48838768e46b94217965185dbb"} Dec 03 16:22:55 crc kubenswrapper[4998]: I1203 16:22:55.902745 4998 generic.go:334] "Generic (PLEG): container finished" podID="72129c14-7e4f-4e0a-9053-7d2178a835ae" containerID="f5cf4c1c2b2ded19a3dd250a7981ea172d494c29024b0caffeef55b427f76d43" exitCode=0 Dec 03 16:22:55 crc kubenswrapper[4998]: I1203 16:22:55.902869 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-fffa-account-create-update-bh4qr" event={"ID":"72129c14-7e4f-4e0a-9053-7d2178a835ae","Type":"ContainerDied","Data":"f5cf4c1c2b2ded19a3dd250a7981ea172d494c29024b0caffeef55b427f76d43"} Dec 03 16:22:55 crc kubenswrapper[4998]: I1203 16:22:55.905622 4998 generic.go:334] "Generic (PLEG): container finished" podID="3739b197-1e59-48fa-a7f7-9e73b95a382a" containerID="88950ec72a9b62516073acb253b694726d83368ce8222ef79a223454a9bba802" exitCode=0 Dec 03 16:22:55 crc kubenswrapper[4998]: I1203 16:22:55.905700 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-p4448" event={"ID":"3739b197-1e59-48fa-a7f7-9e73b95a382a","Type":"ContainerDied","Data":"88950ec72a9b62516073acb253b694726d83368ce8222ef79a223454a9bba802"} Dec 03 16:22:55 crc kubenswrapper[4998]: I1203 16:22:55.907894 4998 generic.go:334] "Generic (PLEG): container finished" podID="0ce04863-7333-4484-9507-93143d276e47" containerID="9fe611c4100838470d9ddd83b84950c87202cecce825b8bc1f19e188b93071e6" exitCode=0 Dec 03 16:22:55 crc kubenswrapper[4998]: I1203 16:22:55.907975 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gwldb" event={"ID":"0ce04863-7333-4484-9507-93143d276e47","Type":"ContainerDied","Data":"9fe611c4100838470d9ddd83b84950c87202cecce825b8bc1f19e188b93071e6"} Dec 03 16:22:55 crc kubenswrapper[4998]: I1203 16:22:55.916630 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-21ed-account-create-update-qnv6c" podStartSLOduration=9.916610089 podStartE2EDuration="9.916610089s" podCreationTimestamp="2025-12-03 16:22:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:22:55.914603729 +0000 UTC m=+1154.526303972" watchObservedRunningTime="2025-12-03 16:22:55.916610089 +0000 UTC m=+1154.528310312" Dec 03 16:22:56 crc kubenswrapper[4998]: I1203 16:22:56.936698 4998 generic.go:334] "Generic (PLEG): container finished" podID="15941ab6-753a-4eb7-8e6f-51ffae3a0227" containerID="75385576e42fd8cb783f5ce0d82d0c0520f8d6a3e21af41b30c90d912cd5f3b0" exitCode=0 Dec 03 16:22:56 crc kubenswrapper[4998]: I1203 16:22:56.936785 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-21ed-account-create-update-qnv6c" event={"ID":"15941ab6-753a-4eb7-8e6f-51ffae3a0227","Type":"ContainerDied","Data":"75385576e42fd8cb783f5ce0d82d0c0520f8d6a3e21af41b30c90d912cd5f3b0"} Dec 03 16:22:56 crc kubenswrapper[4998]: I1203 16:22:56.941769 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92d2d9f9-b0c3-40f8-981a-41ec47f030e1","Type":"ContainerStarted","Data":"392644aa28c6eff25fa63b88178eef2345ac2185aecdb1961b66fdaee4d65dcc"} Dec 03 16:22:57 crc kubenswrapper[4998]: I1203 16:22:57.953399 4998 generic.go:334] "Generic (PLEG): container finished" podID="dbde2166-b545-4979-845a-8932845b9cd1" containerID="5ab3b6bc5babd3b03187dd12c421440a4984c2949d620b458c03048793037989" exitCode=0 Dec 03 16:22:57 crc kubenswrapper[4998]: I1203 16:22:57.953496 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"dbde2166-b545-4979-845a-8932845b9cd1","Type":"ContainerDied","Data":"5ab3b6bc5babd3b03187dd12c421440a4984c2949d620b458c03048793037989"} Dec 03 16:22:57 crc kubenswrapper[4998]: I1203 16:22:57.963281 4998 generic.go:334] "Generic (PLEG): container finished" podID="aeb4a971-3c03-457c-b0df-1503701aac5b" containerID="d9927260898b6eed6b91ad5a3c179db33be6ad3449ab3cb52a37f8296becfaed" exitCode=0 Dec 03 16:22:57 crc kubenswrapper[4998]: I1203 16:22:57.963352 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"aeb4a971-3c03-457c-b0df-1503701aac5b","Type":"ContainerDied","Data":"d9927260898b6eed6b91ad5a3c179db33be6ad3449ab3cb52a37f8296becfaed"} Dec 03 16:22:57 crc kubenswrapper[4998]: I1203 16:22:57.966556 4998 generic.go:334] "Generic (PLEG): container finished" podID="e0ef6669-5a8b-462d-9681-925859c86df1" containerID="a3f53d09f4c25f953ec5e8b2e0a067502b460ce0068895898828046c063bc5c9" exitCode=0 Dec 03 16:22:57 crc kubenswrapper[4998]: I1203 16:22:57.966705 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e0ef6669-5a8b-462d-9681-925859c86df1","Type":"ContainerDied","Data":"a3f53d09f4c25f953ec5e8b2e0a067502b460ce0068895898828046c063bc5c9"} Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.454619 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-fffa-account-create-update-bh4qr" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.463571 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-daeb-account-create-update-vsrjb" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.497155 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-p4448" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.542390 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-r4skm" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.550294 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gwldb" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.561603 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3739b197-1e59-48fa-a7f7-9e73b95a382a-operator-scripts\") pod \"3739b197-1e59-48fa-a7f7-9e73b95a382a\" (UID: \"3739b197-1e59-48fa-a7f7-9e73b95a382a\") " Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.561652 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b542f7b1-ca06-4b56-b45c-dccde2fca019-operator-scripts\") pod \"b542f7b1-ca06-4b56-b45c-dccde2fca019\" (UID: \"b542f7b1-ca06-4b56-b45c-dccde2fca019\") " Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.561680 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s27v9\" (UniqueName: \"kubernetes.io/projected/72129c14-7e4f-4e0a-9053-7d2178a835ae-kube-api-access-s27v9\") pod \"72129c14-7e4f-4e0a-9053-7d2178a835ae\" (UID: \"72129c14-7e4f-4e0a-9053-7d2178a835ae\") " Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.561711 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zjpz\" (UniqueName: \"kubernetes.io/projected/3739b197-1e59-48fa-a7f7-9e73b95a382a-kube-api-access-4zjpz\") pod \"3739b197-1e59-48fa-a7f7-9e73b95a382a\" (UID: \"3739b197-1e59-48fa-a7f7-9e73b95a382a\") " Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.561762 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72129c14-7e4f-4e0a-9053-7d2178a835ae-operator-scripts\") pod \"72129c14-7e4f-4e0a-9053-7d2178a835ae\" (UID: \"72129c14-7e4f-4e0a-9053-7d2178a835ae\") " Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.561805 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqfxt\" (UniqueName: \"kubernetes.io/projected/b542f7b1-ca06-4b56-b45c-dccde2fca019-kube-api-access-nqfxt\") pod \"b542f7b1-ca06-4b56-b45c-dccde2fca019\" (UID: \"b542f7b1-ca06-4b56-b45c-dccde2fca019\") " Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.566491 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72129c14-7e4f-4e0a-9053-7d2178a835ae-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "72129c14-7e4f-4e0a-9053-7d2178a835ae" (UID: "72129c14-7e4f-4e0a-9053-7d2178a835ae"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.566797 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b542f7b1-ca06-4b56-b45c-dccde2fca019-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b542f7b1-ca06-4b56-b45c-dccde2fca019" (UID: "b542f7b1-ca06-4b56-b45c-dccde2fca019"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.566844 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3739b197-1e59-48fa-a7f7-9e73b95a382a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3739b197-1e59-48fa-a7f7-9e73b95a382a" (UID: "3739b197-1e59-48fa-a7f7-9e73b95a382a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.568699 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b542f7b1-ca06-4b56-b45c-dccde2fca019-kube-api-access-nqfxt" (OuterVolumeSpecName: "kube-api-access-nqfxt") pod "b542f7b1-ca06-4b56-b45c-dccde2fca019" (UID: "b542f7b1-ca06-4b56-b45c-dccde2fca019"). InnerVolumeSpecName "kube-api-access-nqfxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.569550 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72129c14-7e4f-4e0a-9053-7d2178a835ae-kube-api-access-s27v9" (OuterVolumeSpecName: "kube-api-access-s27v9") pod "72129c14-7e4f-4e0a-9053-7d2178a835ae" (UID: "72129c14-7e4f-4e0a-9053-7d2178a835ae"). InnerVolumeSpecName "kube-api-access-s27v9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.569870 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3739b197-1e59-48fa-a7f7-9e73b95a382a-kube-api-access-4zjpz" (OuterVolumeSpecName: "kube-api-access-4zjpz") pod "3739b197-1e59-48fa-a7f7-9e73b95a382a" (UID: "3739b197-1e59-48fa-a7f7-9e73b95a382a"). InnerVolumeSpecName "kube-api-access-4zjpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.573343 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-21ed-account-create-update-qnv6c" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.663291 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d71c5697-f3d2-4c29-a722-49823fa55d6d-operator-scripts\") pod \"d71c5697-f3d2-4c29-a722-49823fa55d6d\" (UID: \"d71c5697-f3d2-4c29-a722-49823fa55d6d\") " Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.663419 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lldh\" (UniqueName: \"kubernetes.io/projected/0ce04863-7333-4484-9507-93143d276e47-kube-api-access-2lldh\") pod \"0ce04863-7333-4484-9507-93143d276e47\" (UID: \"0ce04863-7333-4484-9507-93143d276e47\") " Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.663452 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ce04863-7333-4484-9507-93143d276e47-operator-scripts\") pod \"0ce04863-7333-4484-9507-93143d276e47\" (UID: \"0ce04863-7333-4484-9507-93143d276e47\") " Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.663493 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48d7v\" (UniqueName: \"kubernetes.io/projected/d71c5697-f3d2-4c29-a722-49823fa55d6d-kube-api-access-48d7v\") pod \"d71c5697-f3d2-4c29-a722-49823fa55d6d\" (UID: \"d71c5697-f3d2-4c29-a722-49823fa55d6d\") " Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.663526 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ml7z\" (UniqueName: \"kubernetes.io/projected/15941ab6-753a-4eb7-8e6f-51ffae3a0227-kube-api-access-8ml7z\") pod \"15941ab6-753a-4eb7-8e6f-51ffae3a0227\" (UID: \"15941ab6-753a-4eb7-8e6f-51ffae3a0227\") " Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.663591 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15941ab6-753a-4eb7-8e6f-51ffae3a0227-operator-scripts\") pod \"15941ab6-753a-4eb7-8e6f-51ffae3a0227\" (UID: \"15941ab6-753a-4eb7-8e6f-51ffae3a0227\") " Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.663920 4998 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3739b197-1e59-48fa-a7f7-9e73b95a382a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.663941 4998 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b542f7b1-ca06-4b56-b45c-dccde2fca019-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.663954 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s27v9\" (UniqueName: \"kubernetes.io/projected/72129c14-7e4f-4e0a-9053-7d2178a835ae-kube-api-access-s27v9\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.663966 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zjpz\" (UniqueName: \"kubernetes.io/projected/3739b197-1e59-48fa-a7f7-9e73b95a382a-kube-api-access-4zjpz\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.663978 4998 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/72129c14-7e4f-4e0a-9053-7d2178a835ae-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.663989 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqfxt\" (UniqueName: \"kubernetes.io/projected/b542f7b1-ca06-4b56-b45c-dccde2fca019-kube-api-access-nqfxt\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.664325 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15941ab6-753a-4eb7-8e6f-51ffae3a0227-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "15941ab6-753a-4eb7-8e6f-51ffae3a0227" (UID: "15941ab6-753a-4eb7-8e6f-51ffae3a0227"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.664658 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d71c5697-f3d2-4c29-a722-49823fa55d6d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d71c5697-f3d2-4c29-a722-49823fa55d6d" (UID: "d71c5697-f3d2-4c29-a722-49823fa55d6d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.665507 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ce04863-7333-4484-9507-93143d276e47-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0ce04863-7333-4484-9507-93143d276e47" (UID: "0ce04863-7333-4484-9507-93143d276e47"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.667639 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ce04863-7333-4484-9507-93143d276e47-kube-api-access-2lldh" (OuterVolumeSpecName: "kube-api-access-2lldh") pod "0ce04863-7333-4484-9507-93143d276e47" (UID: "0ce04863-7333-4484-9507-93143d276e47"). InnerVolumeSpecName "kube-api-access-2lldh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.668917 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15941ab6-753a-4eb7-8e6f-51ffae3a0227-kube-api-access-8ml7z" (OuterVolumeSpecName: "kube-api-access-8ml7z") pod "15941ab6-753a-4eb7-8e6f-51ffae3a0227" (UID: "15941ab6-753a-4eb7-8e6f-51ffae3a0227"). InnerVolumeSpecName "kube-api-access-8ml7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.668969 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d71c5697-f3d2-4c29-a722-49823fa55d6d-kube-api-access-48d7v" (OuterVolumeSpecName: "kube-api-access-48d7v") pod "d71c5697-f3d2-4c29-a722-49823fa55d6d" (UID: "d71c5697-f3d2-4c29-a722-49823fa55d6d"). InnerVolumeSpecName "kube-api-access-48d7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.767090 4998 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15941ab6-753a-4eb7-8e6f-51ffae3a0227-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.767118 4998 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d71c5697-f3d2-4c29-a722-49823fa55d6d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.767130 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lldh\" (UniqueName: \"kubernetes.io/projected/0ce04863-7333-4484-9507-93143d276e47-kube-api-access-2lldh\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.767139 4998 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ce04863-7333-4484-9507-93143d276e47-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.767149 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48d7v\" (UniqueName: \"kubernetes.io/projected/d71c5697-f3d2-4c29-a722-49823fa55d6d-kube-api-access-48d7v\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.767158 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ml7z\" (UniqueName: \"kubernetes.io/projected/15941ab6-753a-4eb7-8e6f-51ffae3a0227-kube-api-access-8ml7z\") on node \"crc\" DevicePath \"\"" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.982030 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-notifications-server-0" event={"ID":"dbde2166-b545-4979-845a-8932845b9cd1","Type":"ContainerStarted","Data":"8ab031f20b9f6bd9446fd1d905faa668cd2e00d1f776e405b5c8cea0fe34ecb8"} Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.983562 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.987812 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"aeb4a971-3c03-457c-b0df-1503701aac5b","Type":"ContainerStarted","Data":"26aefd457134bd65e1e106463be93313c5b1fefdf9a1b270d124f5ef4d5ce405"} Dec 03 16:22:58 crc kubenswrapper[4998]: I1203 16:22:58.988032 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 16:22:59 crc kubenswrapper[4998]: I1203 16:22:59.008487 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-fffa-account-create-update-bh4qr" event={"ID":"72129c14-7e4f-4e0a-9053-7d2178a835ae","Type":"ContainerDied","Data":"70d9067c3c9c8379cb947632219ffc0bcd05917396ad73334e0da048bb2fd8f4"} Dec 03 16:22:59 crc kubenswrapper[4998]: I1203 16:22:59.008512 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-fffa-account-create-update-bh4qr" Dec 03 16:22:59 crc kubenswrapper[4998]: I1203 16:22:59.008527 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70d9067c3c9c8379cb947632219ffc0bcd05917396ad73334e0da048bb2fd8f4" Dec 03 16:22:59 crc kubenswrapper[4998]: I1203 16:22:59.010569 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-p4448" event={"ID":"3739b197-1e59-48fa-a7f7-9e73b95a382a","Type":"ContainerDied","Data":"81b6bcecbb278d74851fe2744b5be1b0c56a1106d4418c169feb33670878badc"} Dec 03 16:22:59 crc kubenswrapper[4998]: I1203 16:22:59.010599 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81b6bcecbb278d74851fe2744b5be1b0c56a1106d4418c169feb33670878badc" Dec 03 16:22:59 crc kubenswrapper[4998]: I1203 16:22:59.010583 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-p4448" Dec 03 16:22:59 crc kubenswrapper[4998]: I1203 16:22:59.012257 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gwldb" Dec 03 16:22:59 crc kubenswrapper[4998]: I1203 16:22:59.012331 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gwldb" event={"ID":"0ce04863-7333-4484-9507-93143d276e47","Type":"ContainerDied","Data":"566908bc977b146ab72ce9e307b7202dc74e0053df3769c5a7dbd6f1b3329fb7"} Dec 03 16:22:59 crc kubenswrapper[4998]: I1203 16:22:59.012366 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="566908bc977b146ab72ce9e307b7202dc74e0053df3769c5a7dbd6f1b3329fb7" Dec 03 16:22:59 crc kubenswrapper[4998]: I1203 16:22:59.015867 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-notifications-server-0" podStartSLOduration=59.732813257 podStartE2EDuration="1m10.015851409s" podCreationTimestamp="2025-12-03 16:21:49 +0000 UTC" firstStartedPulling="2025-12-03 16:22:12.238627402 +0000 UTC m=+1110.850327625" lastFinishedPulling="2025-12-03 16:22:22.521665514 +0000 UTC m=+1121.133365777" observedRunningTime="2025-12-03 16:22:59.004444712 +0000 UTC m=+1157.616144935" watchObservedRunningTime="2025-12-03 16:22:59.015851409 +0000 UTC m=+1157.627551632" Dec 03 16:22:59 crc kubenswrapper[4998]: I1203 16:22:59.018978 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-daeb-account-create-update-vsrjb" Dec 03 16:22:59 crc kubenswrapper[4998]: I1203 16:22:59.019035 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-daeb-account-create-update-vsrjb" event={"ID":"b542f7b1-ca06-4b56-b45c-dccde2fca019","Type":"ContainerDied","Data":"501ae843f2c6c45ed3119d58a1e93d5c2428b59ad584d62b13128c74cd647922"} Dec 03 16:22:59 crc kubenswrapper[4998]: I1203 16:22:59.019081 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="501ae843f2c6c45ed3119d58a1e93d5c2428b59ad584d62b13128c74cd647922" Dec 03 16:22:59 crc kubenswrapper[4998]: I1203 16:22:59.025068 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-xn4z8" event={"ID":"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c","Type":"ContainerStarted","Data":"6b9ccdd824d76760d64588cbaf85f2f172e2d0f7f41007a34c67b6c6dfa558da"} Dec 03 16:22:59 crc kubenswrapper[4998]: I1203 16:22:59.027342 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-21ed-account-create-update-qnv6c" event={"ID":"15941ab6-753a-4eb7-8e6f-51ffae3a0227","Type":"ContainerDied","Data":"ee3391e8727875e73a5391f7bd33ffa15a1459a3bcc003cca76a9d40630a9a4e"} Dec 03 16:22:59 crc kubenswrapper[4998]: I1203 16:22:59.027365 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee3391e8727875e73a5391f7bd33ffa15a1459a3bcc003cca76a9d40630a9a4e" Dec 03 16:22:59 crc kubenswrapper[4998]: I1203 16:22:59.027400 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-21ed-account-create-update-qnv6c" Dec 03 16:22:59 crc kubenswrapper[4998]: I1203 16:22:59.031181 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-r4skm" event={"ID":"d71c5697-f3d2-4c29-a722-49823fa55d6d","Type":"ContainerDied","Data":"8a3af9fd80215e1f279d7af6e8b800d5b85aa43c98e32d07d842ef59cd811258"} Dec 03 16:22:59 crc kubenswrapper[4998]: I1203 16:22:59.031206 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a3af9fd80215e1f279d7af6e8b800d5b85aa43c98e32d07d842ef59cd811258" Dec 03 16:22:59 crc kubenswrapper[4998]: I1203 16:22:59.031243 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-r4skm" Dec 03 16:22:59 crc kubenswrapper[4998]: I1203 16:22:59.036935 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e0ef6669-5a8b-462d-9681-925859c86df1","Type":"ContainerStarted","Data":"1a47cb6169680516ff6125777af3aa66a1003866aa63ca4fe29fe2f67207bf14"} Dec 03 16:22:59 crc kubenswrapper[4998]: I1203 16:22:59.037869 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:22:59 crc kubenswrapper[4998]: I1203 16:22:59.061414 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=59.602379878 podStartE2EDuration="1m11.061398267s" podCreationTimestamp="2025-12-03 16:21:48 +0000 UTC" firstStartedPulling="2025-12-03 16:22:11.062600133 +0000 UTC m=+1109.674300356" lastFinishedPulling="2025-12-03 16:22:22.521618482 +0000 UTC m=+1121.133318745" observedRunningTime="2025-12-03 16:22:59.049880296 +0000 UTC m=+1157.661580519" watchObservedRunningTime="2025-12-03 16:22:59.061398267 +0000 UTC m=+1157.673098490" Dec 03 16:22:59 crc kubenswrapper[4998]: I1203 16:22:59.089321 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=59.416697653 podStartE2EDuration="1m10.089300079s" podCreationTimestamp="2025-12-03 16:21:49 +0000 UTC" firstStartedPulling="2025-12-03 16:22:11.10095345 +0000 UTC m=+1109.712653673" lastFinishedPulling="2025-12-03 16:22:21.773555846 +0000 UTC m=+1120.385256099" observedRunningTime="2025-12-03 16:22:59.081443672 +0000 UTC m=+1157.693143895" watchObservedRunningTime="2025-12-03 16:22:59.089300079 +0000 UTC m=+1157.701000302" Dec 03 16:22:59 crc kubenswrapper[4998]: E1203 16:22:59.102164 4998 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb542f7b1_ca06_4b56_b45c_dccde2fca019.slice\": RecentStats: unable to find data in memory cache]" Dec 03 16:22:59 crc kubenswrapper[4998]: I1203 16:22:59.119186 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-xn4z8" podStartSLOduration=13.948740135 podStartE2EDuration="18.119166802s" podCreationTimestamp="2025-12-03 16:22:41 +0000 UTC" firstStartedPulling="2025-12-03 16:22:54.159707717 +0000 UTC m=+1152.771407940" lastFinishedPulling="2025-12-03 16:22:58.330134384 +0000 UTC m=+1156.941834607" observedRunningTime="2025-12-03 16:22:59.105763784 +0000 UTC m=+1157.717464007" watchObservedRunningTime="2025-12-03 16:22:59.119166802 +0000 UTC m=+1157.730867035" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.035367 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-7sxbw" podUID="95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f" containerName="ovn-controller" probeResult="failure" output=< Dec 03 16:23:00 crc kubenswrapper[4998]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 16:23:00 crc kubenswrapper[4998]: > Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.077626 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-5hxf2" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.082125 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-5hxf2" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.342205 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-7sxbw-config-lzbnd"] Dec 03 16:23:00 crc kubenswrapper[4998]: E1203 16:23:00.342600 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ce04863-7333-4484-9507-93143d276e47" containerName="mariadb-database-create" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.342616 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ce04863-7333-4484-9507-93143d276e47" containerName="mariadb-database-create" Dec 03 16:23:00 crc kubenswrapper[4998]: E1203 16:23:00.342643 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15941ab6-753a-4eb7-8e6f-51ffae3a0227" containerName="mariadb-account-create-update" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.342651 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="15941ab6-753a-4eb7-8e6f-51ffae3a0227" containerName="mariadb-account-create-update" Dec 03 16:23:00 crc kubenswrapper[4998]: E1203 16:23:00.342663 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b542f7b1-ca06-4b56-b45c-dccde2fca019" containerName="mariadb-account-create-update" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.342672 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="b542f7b1-ca06-4b56-b45c-dccde2fca019" containerName="mariadb-account-create-update" Dec 03 16:23:00 crc kubenswrapper[4998]: E1203 16:23:00.342688 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d71c5697-f3d2-4c29-a722-49823fa55d6d" containerName="mariadb-database-create" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.342695 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="d71c5697-f3d2-4c29-a722-49823fa55d6d" containerName="mariadb-database-create" Dec 03 16:23:00 crc kubenswrapper[4998]: E1203 16:23:00.342707 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3739b197-1e59-48fa-a7f7-9e73b95a382a" containerName="mariadb-database-create" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.342716 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="3739b197-1e59-48fa-a7f7-9e73b95a382a" containerName="mariadb-database-create" Dec 03 16:23:00 crc kubenswrapper[4998]: E1203 16:23:00.342734 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72129c14-7e4f-4e0a-9053-7d2178a835ae" containerName="mariadb-account-create-update" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.342742 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="72129c14-7e4f-4e0a-9053-7d2178a835ae" containerName="mariadb-account-create-update" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.343005 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="3739b197-1e59-48fa-a7f7-9e73b95a382a" containerName="mariadb-database-create" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.343043 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="d71c5697-f3d2-4c29-a722-49823fa55d6d" containerName="mariadb-database-create" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.343059 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="b542f7b1-ca06-4b56-b45c-dccde2fca019" containerName="mariadb-account-create-update" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.343072 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ce04863-7333-4484-9507-93143d276e47" containerName="mariadb-database-create" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.343083 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="15941ab6-753a-4eb7-8e6f-51ffae3a0227" containerName="mariadb-account-create-update" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.343095 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="72129c14-7e4f-4e0a-9053-7d2178a835ae" containerName="mariadb-account-create-update" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.345930 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7sxbw-config-lzbnd" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.364793 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-7sxbw-config-lzbnd"] Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.376342 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.494293 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f4363fba-b4ac-436f-8816-139522460a9a-additional-scripts\") pod \"ovn-controller-7sxbw-config-lzbnd\" (UID: \"f4363fba-b4ac-436f-8816-139522460a9a\") " pod="openstack/ovn-controller-7sxbw-config-lzbnd" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.494579 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f4363fba-b4ac-436f-8816-139522460a9a-var-run\") pod \"ovn-controller-7sxbw-config-lzbnd\" (UID: \"f4363fba-b4ac-436f-8816-139522460a9a\") " pod="openstack/ovn-controller-7sxbw-config-lzbnd" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.494689 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbmxn\" (UniqueName: \"kubernetes.io/projected/f4363fba-b4ac-436f-8816-139522460a9a-kube-api-access-vbmxn\") pod \"ovn-controller-7sxbw-config-lzbnd\" (UID: \"f4363fba-b4ac-436f-8816-139522460a9a\") " pod="openstack/ovn-controller-7sxbw-config-lzbnd" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.494869 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f4363fba-b4ac-436f-8816-139522460a9a-var-log-ovn\") pod \"ovn-controller-7sxbw-config-lzbnd\" (UID: \"f4363fba-b4ac-436f-8816-139522460a9a\") " pod="openstack/ovn-controller-7sxbw-config-lzbnd" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.494991 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4363fba-b4ac-436f-8816-139522460a9a-scripts\") pod \"ovn-controller-7sxbw-config-lzbnd\" (UID: \"f4363fba-b4ac-436f-8816-139522460a9a\") " pod="openstack/ovn-controller-7sxbw-config-lzbnd" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.495137 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f4363fba-b4ac-436f-8816-139522460a9a-var-run-ovn\") pod \"ovn-controller-7sxbw-config-lzbnd\" (UID: \"f4363fba-b4ac-436f-8816-139522460a9a\") " pod="openstack/ovn-controller-7sxbw-config-lzbnd" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.596957 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f4363fba-b4ac-436f-8816-139522460a9a-var-run\") pod \"ovn-controller-7sxbw-config-lzbnd\" (UID: \"f4363fba-b4ac-436f-8816-139522460a9a\") " pod="openstack/ovn-controller-7sxbw-config-lzbnd" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.597015 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbmxn\" (UniqueName: \"kubernetes.io/projected/f4363fba-b4ac-436f-8816-139522460a9a-kube-api-access-vbmxn\") pod \"ovn-controller-7sxbw-config-lzbnd\" (UID: \"f4363fba-b4ac-436f-8816-139522460a9a\") " pod="openstack/ovn-controller-7sxbw-config-lzbnd" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.597075 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f4363fba-b4ac-436f-8816-139522460a9a-var-log-ovn\") pod \"ovn-controller-7sxbw-config-lzbnd\" (UID: \"f4363fba-b4ac-436f-8816-139522460a9a\") " pod="openstack/ovn-controller-7sxbw-config-lzbnd" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.597098 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4363fba-b4ac-436f-8816-139522460a9a-scripts\") pod \"ovn-controller-7sxbw-config-lzbnd\" (UID: \"f4363fba-b4ac-436f-8816-139522460a9a\") " pod="openstack/ovn-controller-7sxbw-config-lzbnd" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.597140 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f4363fba-b4ac-436f-8816-139522460a9a-var-run-ovn\") pod \"ovn-controller-7sxbw-config-lzbnd\" (UID: \"f4363fba-b4ac-436f-8816-139522460a9a\") " pod="openstack/ovn-controller-7sxbw-config-lzbnd" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.597213 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f4363fba-b4ac-436f-8816-139522460a9a-additional-scripts\") pod \"ovn-controller-7sxbw-config-lzbnd\" (UID: \"f4363fba-b4ac-436f-8816-139522460a9a\") " pod="openstack/ovn-controller-7sxbw-config-lzbnd" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.597386 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f4363fba-b4ac-436f-8816-139522460a9a-var-run-ovn\") pod \"ovn-controller-7sxbw-config-lzbnd\" (UID: \"f4363fba-b4ac-436f-8816-139522460a9a\") " pod="openstack/ovn-controller-7sxbw-config-lzbnd" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.597402 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f4363fba-b4ac-436f-8816-139522460a9a-var-log-ovn\") pod \"ovn-controller-7sxbw-config-lzbnd\" (UID: \"f4363fba-b4ac-436f-8816-139522460a9a\") " pod="openstack/ovn-controller-7sxbw-config-lzbnd" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.597467 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f4363fba-b4ac-436f-8816-139522460a9a-var-run\") pod \"ovn-controller-7sxbw-config-lzbnd\" (UID: \"f4363fba-b4ac-436f-8816-139522460a9a\") " pod="openstack/ovn-controller-7sxbw-config-lzbnd" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.640911 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f4363fba-b4ac-436f-8816-139522460a9a-additional-scripts\") pod \"ovn-controller-7sxbw-config-lzbnd\" (UID: \"f4363fba-b4ac-436f-8816-139522460a9a\") " pod="openstack/ovn-controller-7sxbw-config-lzbnd" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.642430 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4363fba-b4ac-436f-8816-139522460a9a-scripts\") pod \"ovn-controller-7sxbw-config-lzbnd\" (UID: \"f4363fba-b4ac-436f-8816-139522460a9a\") " pod="openstack/ovn-controller-7sxbw-config-lzbnd" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.654744 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbmxn\" (UniqueName: \"kubernetes.io/projected/f4363fba-b4ac-436f-8816-139522460a9a-kube-api-access-vbmxn\") pod \"ovn-controller-7sxbw-config-lzbnd\" (UID: \"f4363fba-b4ac-436f-8816-139522460a9a\") " pod="openstack/ovn-controller-7sxbw-config-lzbnd" Dec 03 16:23:00 crc kubenswrapper[4998]: I1203 16:23:00.707091 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7sxbw-config-lzbnd" Dec 03 16:23:01 crc kubenswrapper[4998]: I1203 16:23:01.161580 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-7sxbw-config-lzbnd"] Dec 03 16:23:01 crc kubenswrapper[4998]: W1203 16:23:01.167989 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4363fba_b4ac_436f_8816_139522460a9a.slice/crio-23b32d198d89f10228082e4e23fbd229fdda6fc953ec49bed24727b2c9aa557b WatchSource:0}: Error finding container 23b32d198d89f10228082e4e23fbd229fdda6fc953ec49bed24727b2c9aa557b: Status 404 returned error can't find the container with id 23b32d198d89f10228082e4e23fbd229fdda6fc953ec49bed24727b2c9aa557b Dec 03 16:23:02 crc kubenswrapper[4998]: I1203 16:23:02.069800 4998 generic.go:334] "Generic (PLEG): container finished" podID="f4363fba-b4ac-436f-8816-139522460a9a" containerID="b88c2ed698f351d842f6a803ae1df5e84039dc05b78d6abfdee9e43bf12ecd18" exitCode=0 Dec 03 16:23:02 crc kubenswrapper[4998]: I1203 16:23:02.070338 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7sxbw-config-lzbnd" event={"ID":"f4363fba-b4ac-436f-8816-139522460a9a","Type":"ContainerDied","Data":"b88c2ed698f351d842f6a803ae1df5e84039dc05b78d6abfdee9e43bf12ecd18"} Dec 03 16:23:02 crc kubenswrapper[4998]: I1203 16:23:02.070365 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7sxbw-config-lzbnd" event={"ID":"f4363fba-b4ac-436f-8816-139522460a9a","Type":"ContainerStarted","Data":"23b32d198d89f10228082e4e23fbd229fdda6fc953ec49bed24727b2c9aa557b"} Dec 03 16:23:02 crc kubenswrapper[4998]: I1203 16:23:02.136931 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" Dec 03 16:23:02 crc kubenswrapper[4998]: I1203 16:23:02.186498 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b6fcbcc87-jjnql"] Dec 03 16:23:02 crc kubenswrapper[4998]: I1203 16:23:02.186712 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b6fcbcc87-jjnql" podUID="4c6b2505-697f-4935-af28-88a4ae94524f" containerName="dnsmasq-dns" containerID="cri-o://18584dc50961546fa86c00b8e71d0469c200cf58ca1878e37a80d8b933964890" gracePeriod=10 Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.023576 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b6fcbcc87-jjnql" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.081658 4998 generic.go:334] "Generic (PLEG): container finished" podID="4c6b2505-697f-4935-af28-88a4ae94524f" containerID="18584dc50961546fa86c00b8e71d0469c200cf58ca1878e37a80d8b933964890" exitCode=0 Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.081720 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6fcbcc87-jjnql" event={"ID":"4c6b2505-697f-4935-af28-88a4ae94524f","Type":"ContainerDied","Data":"18584dc50961546fa86c00b8e71d0469c200cf58ca1878e37a80d8b933964890"} Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.081748 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6fcbcc87-jjnql" event={"ID":"4c6b2505-697f-4935-af28-88a4ae94524f","Type":"ContainerDied","Data":"73f392ef5b749f98e6bee415211b68f67336f6059571034e2d1c4e63802db5e6"} Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.081769 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b6fcbcc87-jjnql" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.081779 4998 scope.go:117] "RemoveContainer" containerID="18584dc50961546fa86c00b8e71d0469c200cf58ca1878e37a80d8b933964890" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.085414 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92d2d9f9-b0c3-40f8-981a-41ec47f030e1","Type":"ContainerStarted","Data":"99c43f95cd03544481cc3d9f5eefe61127fd464b7135edd8f2e2587a46506ef9"} Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.100466 4998 scope.go:117] "RemoveContainer" containerID="47cdf182c8e1aec57071b2479434e6a1905d6565e7dc73e1564b3e2f9b833d55" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.132108 4998 scope.go:117] "RemoveContainer" containerID="18584dc50961546fa86c00b8e71d0469c200cf58ca1878e37a80d8b933964890" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.132308 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=16.536068685 podStartE2EDuration="1m7.132290265s" podCreationTimestamp="2025-12-03 16:21:56 +0000 UTC" firstStartedPulling="2025-12-03 16:22:12.18378981 +0000 UTC m=+1110.795490033" lastFinishedPulling="2025-12-03 16:23:02.78001139 +0000 UTC m=+1161.391711613" observedRunningTime="2025-12-03 16:23:03.127517735 +0000 UTC m=+1161.739217958" watchObservedRunningTime="2025-12-03 16:23:03.132290265 +0000 UTC m=+1161.743990478" Dec 03 16:23:03 crc kubenswrapper[4998]: E1203 16:23:03.132549 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18584dc50961546fa86c00b8e71d0469c200cf58ca1878e37a80d8b933964890\": container with ID starting with 18584dc50961546fa86c00b8e71d0469c200cf58ca1878e37a80d8b933964890 not found: ID does not exist" containerID="18584dc50961546fa86c00b8e71d0469c200cf58ca1878e37a80d8b933964890" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.132590 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18584dc50961546fa86c00b8e71d0469c200cf58ca1878e37a80d8b933964890"} err="failed to get container status \"18584dc50961546fa86c00b8e71d0469c200cf58ca1878e37a80d8b933964890\": rpc error: code = NotFound desc = could not find container \"18584dc50961546fa86c00b8e71d0469c200cf58ca1878e37a80d8b933964890\": container with ID starting with 18584dc50961546fa86c00b8e71d0469c200cf58ca1878e37a80d8b933964890 not found: ID does not exist" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.132616 4998 scope.go:117] "RemoveContainer" containerID="47cdf182c8e1aec57071b2479434e6a1905d6565e7dc73e1564b3e2f9b833d55" Dec 03 16:23:03 crc kubenswrapper[4998]: E1203 16:23:03.132840 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47cdf182c8e1aec57071b2479434e6a1905d6565e7dc73e1564b3e2f9b833d55\": container with ID starting with 47cdf182c8e1aec57071b2479434e6a1905d6565e7dc73e1564b3e2f9b833d55 not found: ID does not exist" containerID="47cdf182c8e1aec57071b2479434e6a1905d6565e7dc73e1564b3e2f9b833d55" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.132860 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47cdf182c8e1aec57071b2479434e6a1905d6565e7dc73e1564b3e2f9b833d55"} err="failed to get container status \"47cdf182c8e1aec57071b2479434e6a1905d6565e7dc73e1564b3e2f9b833d55\": rpc error: code = NotFound desc = could not find container \"47cdf182c8e1aec57071b2479434e6a1905d6565e7dc73e1564b3e2f9b833d55\": container with ID starting with 47cdf182c8e1aec57071b2479434e6a1905d6565e7dc73e1564b3e2f9b833d55 not found: ID does not exist" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.144032 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzwzl\" (UniqueName: \"kubernetes.io/projected/4c6b2505-697f-4935-af28-88a4ae94524f-kube-api-access-dzwzl\") pod \"4c6b2505-697f-4935-af28-88a4ae94524f\" (UID: \"4c6b2505-697f-4935-af28-88a4ae94524f\") " Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.144202 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c6b2505-697f-4935-af28-88a4ae94524f-config\") pod \"4c6b2505-697f-4935-af28-88a4ae94524f\" (UID: \"4c6b2505-697f-4935-af28-88a4ae94524f\") " Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.144271 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c6b2505-697f-4935-af28-88a4ae94524f-dns-svc\") pod \"4c6b2505-697f-4935-af28-88a4ae94524f\" (UID: \"4c6b2505-697f-4935-af28-88a4ae94524f\") " Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.144309 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c6b2505-697f-4935-af28-88a4ae94524f-ovsdbserver-nb\") pod \"4c6b2505-697f-4935-af28-88a4ae94524f\" (UID: \"4c6b2505-697f-4935-af28-88a4ae94524f\") " Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.150035 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.150150 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c6b2505-697f-4935-af28-88a4ae94524f-kube-api-access-dzwzl" (OuterVolumeSpecName: "kube-api-access-dzwzl") pod "4c6b2505-697f-4935-af28-88a4ae94524f" (UID: "4c6b2505-697f-4935-af28-88a4ae94524f"). InnerVolumeSpecName "kube-api-access-dzwzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.194524 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c6b2505-697f-4935-af28-88a4ae94524f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4c6b2505-697f-4935-af28-88a4ae94524f" (UID: "4c6b2505-697f-4935-af28-88a4ae94524f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.215858 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c6b2505-697f-4935-af28-88a4ae94524f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4c6b2505-697f-4935-af28-88a4ae94524f" (UID: "4c6b2505-697f-4935-af28-88a4ae94524f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.222494 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c6b2505-697f-4935-af28-88a4ae94524f-config" (OuterVolumeSpecName: "config") pod "4c6b2505-697f-4935-af28-88a4ae94524f" (UID: "4c6b2505-697f-4935-af28-88a4ae94524f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.248850 4998 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c6b2505-697f-4935-af28-88a4ae94524f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.248884 4998 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c6b2505-697f-4935-af28-88a4ae94524f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.248894 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzwzl\" (UniqueName: \"kubernetes.io/projected/4c6b2505-697f-4935-af28-88a4ae94524f-kube-api-access-dzwzl\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.248903 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c6b2505-697f-4935-af28-88a4ae94524f-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.343223 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7sxbw-config-lzbnd" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.411251 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b6fcbcc87-jjnql"] Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.418953 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b6fcbcc87-jjnql"] Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.451698 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbmxn\" (UniqueName: \"kubernetes.io/projected/f4363fba-b4ac-436f-8816-139522460a9a-kube-api-access-vbmxn\") pod \"f4363fba-b4ac-436f-8816-139522460a9a\" (UID: \"f4363fba-b4ac-436f-8816-139522460a9a\") " Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.451835 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f4363fba-b4ac-436f-8816-139522460a9a-var-run\") pod \"f4363fba-b4ac-436f-8816-139522460a9a\" (UID: \"f4363fba-b4ac-436f-8816-139522460a9a\") " Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.451878 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4363fba-b4ac-436f-8816-139522460a9a-var-run" (OuterVolumeSpecName: "var-run") pod "f4363fba-b4ac-436f-8816-139522460a9a" (UID: "f4363fba-b4ac-436f-8816-139522460a9a"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.451923 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4363fba-b4ac-436f-8816-139522460a9a-scripts\") pod \"f4363fba-b4ac-436f-8816-139522460a9a\" (UID: \"f4363fba-b4ac-436f-8816-139522460a9a\") " Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.451946 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f4363fba-b4ac-436f-8816-139522460a9a-var-log-ovn\") pod \"f4363fba-b4ac-436f-8816-139522460a9a\" (UID: \"f4363fba-b4ac-436f-8816-139522460a9a\") " Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.451989 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f4363fba-b4ac-436f-8816-139522460a9a-var-run-ovn\") pod \"f4363fba-b4ac-436f-8816-139522460a9a\" (UID: \"f4363fba-b4ac-436f-8816-139522460a9a\") " Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.452037 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4363fba-b4ac-436f-8816-139522460a9a-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "f4363fba-b4ac-436f-8816-139522460a9a" (UID: "f4363fba-b4ac-436f-8816-139522460a9a"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.452052 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f4363fba-b4ac-436f-8816-139522460a9a-additional-scripts\") pod \"f4363fba-b4ac-436f-8816-139522460a9a\" (UID: \"f4363fba-b4ac-436f-8816-139522460a9a\") " Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.452072 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4363fba-b4ac-436f-8816-139522460a9a-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "f4363fba-b4ac-436f-8816-139522460a9a" (UID: "f4363fba-b4ac-436f-8816-139522460a9a"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.452653 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4363fba-b4ac-436f-8816-139522460a9a-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "f4363fba-b4ac-436f-8816-139522460a9a" (UID: "f4363fba-b4ac-436f-8816-139522460a9a"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.452960 4998 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f4363fba-b4ac-436f-8816-139522460a9a-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.452976 4998 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f4363fba-b4ac-436f-8816-139522460a9a-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.452984 4998 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f4363fba-b4ac-436f-8816-139522460a9a-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.452997 4998 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f4363fba-b4ac-436f-8816-139522460a9a-var-run\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.453000 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4363fba-b4ac-436f-8816-139522460a9a-scripts" (OuterVolumeSpecName: "scripts") pod "f4363fba-b4ac-436f-8816-139522460a9a" (UID: "f4363fba-b4ac-436f-8816-139522460a9a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.454893 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4363fba-b4ac-436f-8816-139522460a9a-kube-api-access-vbmxn" (OuterVolumeSpecName: "kube-api-access-vbmxn") pod "f4363fba-b4ac-436f-8816-139522460a9a" (UID: "f4363fba-b4ac-436f-8816-139522460a9a"). InnerVolumeSpecName "kube-api-access-vbmxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.555051 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4363fba-b4ac-436f-8816-139522460a9a-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.555089 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbmxn\" (UniqueName: \"kubernetes.io/projected/f4363fba-b4ac-436f-8816-139522460a9a-kube-api-access-vbmxn\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:03 crc kubenswrapper[4998]: I1203 16:23:03.688465 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c6b2505-697f-4935-af28-88a4ae94524f" path="/var/lib/kubelet/pods/4c6b2505-697f-4935-af28-88a4ae94524f/volumes" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.098381 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7sxbw-config-lzbnd" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.098379 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7sxbw-config-lzbnd" event={"ID":"f4363fba-b4ac-436f-8816-139522460a9a","Type":"ContainerDied","Data":"23b32d198d89f10228082e4e23fbd229fdda6fc953ec49bed24727b2c9aa557b"} Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.098517 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23b32d198d89f10228082e4e23fbd229fdda6fc953ec49bed24727b2c9aa557b" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.476154 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-7sxbw-config-lzbnd"] Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.490768 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-7sxbw-config-lzbnd"] Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.658520 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-7sxbw-config-l66rt"] Dec 03 16:23:04 crc kubenswrapper[4998]: E1203 16:23:04.658864 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c6b2505-697f-4935-af28-88a4ae94524f" containerName="dnsmasq-dns" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.658880 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c6b2505-697f-4935-af28-88a4ae94524f" containerName="dnsmasq-dns" Dec 03 16:23:04 crc kubenswrapper[4998]: E1203 16:23:04.658894 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4363fba-b4ac-436f-8816-139522460a9a" containerName="ovn-config" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.658900 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4363fba-b4ac-436f-8816-139522460a9a" containerName="ovn-config" Dec 03 16:23:04 crc kubenswrapper[4998]: E1203 16:23:04.658923 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c6b2505-697f-4935-af28-88a4ae94524f" containerName="init" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.658930 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c6b2505-697f-4935-af28-88a4ae94524f" containerName="init" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.659094 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c6b2505-697f-4935-af28-88a4ae94524f" containerName="dnsmasq-dns" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.659123 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4363fba-b4ac-436f-8816-139522460a9a" containerName="ovn-config" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.659669 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7sxbw-config-l66rt" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.662181 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.671312 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-7sxbw-config-l66rt"] Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.774411 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6a1d3e43-409c-48c7-a6ba-dcd77380751e-additional-scripts\") pod \"ovn-controller-7sxbw-config-l66rt\" (UID: \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\") " pod="openstack/ovn-controller-7sxbw-config-l66rt" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.774526 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a1d3e43-409c-48c7-a6ba-dcd77380751e-scripts\") pod \"ovn-controller-7sxbw-config-l66rt\" (UID: \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\") " pod="openstack/ovn-controller-7sxbw-config-l66rt" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.774550 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6a1d3e43-409c-48c7-a6ba-dcd77380751e-var-log-ovn\") pod \"ovn-controller-7sxbw-config-l66rt\" (UID: \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\") " pod="openstack/ovn-controller-7sxbw-config-l66rt" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.774883 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6a1d3e43-409c-48c7-a6ba-dcd77380751e-var-run\") pod \"ovn-controller-7sxbw-config-l66rt\" (UID: \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\") " pod="openstack/ovn-controller-7sxbw-config-l66rt" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.774939 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6a1d3e43-409c-48c7-a6ba-dcd77380751e-var-run-ovn\") pod \"ovn-controller-7sxbw-config-l66rt\" (UID: \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\") " pod="openstack/ovn-controller-7sxbw-config-l66rt" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.774990 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k25fr\" (UniqueName: \"kubernetes.io/projected/6a1d3e43-409c-48c7-a6ba-dcd77380751e-kube-api-access-k25fr\") pod \"ovn-controller-7sxbw-config-l66rt\" (UID: \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\") " pod="openstack/ovn-controller-7sxbw-config-l66rt" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.876964 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6a1d3e43-409c-48c7-a6ba-dcd77380751e-var-run\") pod \"ovn-controller-7sxbw-config-l66rt\" (UID: \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\") " pod="openstack/ovn-controller-7sxbw-config-l66rt" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.877013 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6a1d3e43-409c-48c7-a6ba-dcd77380751e-var-run-ovn\") pod \"ovn-controller-7sxbw-config-l66rt\" (UID: \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\") " pod="openstack/ovn-controller-7sxbw-config-l66rt" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.877035 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k25fr\" (UniqueName: \"kubernetes.io/projected/6a1d3e43-409c-48c7-a6ba-dcd77380751e-kube-api-access-k25fr\") pod \"ovn-controller-7sxbw-config-l66rt\" (UID: \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\") " pod="openstack/ovn-controller-7sxbw-config-l66rt" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.877078 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6a1d3e43-409c-48c7-a6ba-dcd77380751e-additional-scripts\") pod \"ovn-controller-7sxbw-config-l66rt\" (UID: \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\") " pod="openstack/ovn-controller-7sxbw-config-l66rt" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.877112 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a1d3e43-409c-48c7-a6ba-dcd77380751e-scripts\") pod \"ovn-controller-7sxbw-config-l66rt\" (UID: \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\") " pod="openstack/ovn-controller-7sxbw-config-l66rt" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.877127 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6a1d3e43-409c-48c7-a6ba-dcd77380751e-var-log-ovn\") pod \"ovn-controller-7sxbw-config-l66rt\" (UID: \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\") " pod="openstack/ovn-controller-7sxbw-config-l66rt" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.877229 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6a1d3e43-409c-48c7-a6ba-dcd77380751e-var-log-ovn\") pod \"ovn-controller-7sxbw-config-l66rt\" (UID: \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\") " pod="openstack/ovn-controller-7sxbw-config-l66rt" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.877243 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6a1d3e43-409c-48c7-a6ba-dcd77380751e-var-run\") pod \"ovn-controller-7sxbw-config-l66rt\" (UID: \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\") " pod="openstack/ovn-controller-7sxbw-config-l66rt" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.877276 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6a1d3e43-409c-48c7-a6ba-dcd77380751e-var-run-ovn\") pod \"ovn-controller-7sxbw-config-l66rt\" (UID: \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\") " pod="openstack/ovn-controller-7sxbw-config-l66rt" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.877885 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6a1d3e43-409c-48c7-a6ba-dcd77380751e-additional-scripts\") pod \"ovn-controller-7sxbw-config-l66rt\" (UID: \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\") " pod="openstack/ovn-controller-7sxbw-config-l66rt" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.879073 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a1d3e43-409c-48c7-a6ba-dcd77380751e-scripts\") pod \"ovn-controller-7sxbw-config-l66rt\" (UID: \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\") " pod="openstack/ovn-controller-7sxbw-config-l66rt" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.906195 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k25fr\" (UniqueName: \"kubernetes.io/projected/6a1d3e43-409c-48c7-a6ba-dcd77380751e-kube-api-access-k25fr\") pod \"ovn-controller-7sxbw-config-l66rt\" (UID: \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\") " pod="openstack/ovn-controller-7sxbw-config-l66rt" Dec 03 16:23:04 crc kubenswrapper[4998]: I1203 16:23:04.973668 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7sxbw-config-l66rt" Dec 03 16:23:05 crc kubenswrapper[4998]: I1203 16:23:05.046312 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-7sxbw" Dec 03 16:23:05 crc kubenswrapper[4998]: I1203 16:23:05.452545 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-7sxbw-config-l66rt"] Dec 03 16:23:05 crc kubenswrapper[4998]: W1203 16:23:05.455137 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a1d3e43_409c_48c7_a6ba_dcd77380751e.slice/crio-1a2dda1587ae7ba3c1adfeca96cdc488f3afc83114ef84a541db7d56717d1b49 WatchSource:0}: Error finding container 1a2dda1587ae7ba3c1adfeca96cdc488f3afc83114ef84a541db7d56717d1b49: Status 404 returned error can't find the container with id 1a2dda1587ae7ba3c1adfeca96cdc488f3afc83114ef84a541db7d56717d1b49 Dec 03 16:23:05 crc kubenswrapper[4998]: I1203 16:23:05.706268 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4363fba-b4ac-436f-8816-139522460a9a" path="/var/lib/kubelet/pods/f4363fba-b4ac-436f-8816-139522460a9a/volumes" Dec 03 16:23:06 crc kubenswrapper[4998]: I1203 16:23:06.120184 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7sxbw-config-l66rt" event={"ID":"6a1d3e43-409c-48c7-a6ba-dcd77380751e","Type":"ContainerStarted","Data":"1a2dda1587ae7ba3c1adfeca96cdc488f3afc83114ef84a541db7d56717d1b49"} Dec 03 16:23:07 crc kubenswrapper[4998]: I1203 16:23:07.128984 4998 generic.go:334] "Generic (PLEG): container finished" podID="78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c" containerID="6b9ccdd824d76760d64588cbaf85f2f172e2d0f7f41007a34c67b6c6dfa558da" exitCode=0 Dec 03 16:23:07 crc kubenswrapper[4998]: I1203 16:23:07.129055 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-xn4z8" event={"ID":"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c","Type":"ContainerDied","Data":"6b9ccdd824d76760d64588cbaf85f2f172e2d0f7f41007a34c67b6c6dfa558da"} Dec 03 16:23:07 crc kubenswrapper[4998]: I1203 16:23:07.130821 4998 generic.go:334] "Generic (PLEG): container finished" podID="6a1d3e43-409c-48c7-a6ba-dcd77380751e" containerID="ae868c4032d556a63210bdc8ccf9b359af563c9380921b0a12a6d6c67517a6c8" exitCode=0 Dec 03 16:23:07 crc kubenswrapper[4998]: I1203 16:23:07.130872 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7sxbw-config-l66rt" event={"ID":"6a1d3e43-409c-48c7-a6ba-dcd77380751e","Type":"ContainerDied","Data":"ae868c4032d556a63210bdc8ccf9b359af563c9380921b0a12a6d6c67517a6c8"} Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.575884 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7sxbw-config-l66rt" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.582383 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-xn4z8" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.751290 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a1d3e43-409c-48c7-a6ba-dcd77380751e-scripts\") pod \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\" (UID: \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\") " Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.751367 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-swiftconf\") pod \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.751386 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6a1d3e43-409c-48c7-a6ba-dcd77380751e-var-log-ovn\") pod \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\" (UID: \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\") " Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.751420 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-ring-data-devices\") pod \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.751463 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7vj2\" (UniqueName: \"kubernetes.io/projected/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-kube-api-access-q7vj2\") pod \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.751492 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a1d3e43-409c-48c7-a6ba-dcd77380751e-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "6a1d3e43-409c-48c7-a6ba-dcd77380751e" (UID: "6a1d3e43-409c-48c7-a6ba-dcd77380751e"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.751501 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-combined-ca-bundle\") pod \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.751554 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6a1d3e43-409c-48c7-a6ba-dcd77380751e-var-run-ovn\") pod \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\" (UID: \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\") " Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.751593 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-dispersionconf\") pod \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.751694 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6a1d3e43-409c-48c7-a6ba-dcd77380751e-additional-scripts\") pod \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\" (UID: \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\") " Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.751713 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6a1d3e43-409c-48c7-a6ba-dcd77380751e-var-run\") pod \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\" (UID: \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\") " Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.751733 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-scripts\") pod \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.751787 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-etc-swift\") pod \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\" (UID: \"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c\") " Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.751810 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k25fr\" (UniqueName: \"kubernetes.io/projected/6a1d3e43-409c-48c7-a6ba-dcd77380751e-kube-api-access-k25fr\") pod \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\" (UID: \"6a1d3e43-409c-48c7-a6ba-dcd77380751e\") " Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.752084 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a1d3e43-409c-48c7-a6ba-dcd77380751e-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "6a1d3e43-409c-48c7-a6ba-dcd77380751e" (UID: "6a1d3e43-409c-48c7-a6ba-dcd77380751e"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.752202 4998 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6a1d3e43-409c-48c7-a6ba-dcd77380751e-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.752215 4998 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6a1d3e43-409c-48c7-a6ba-dcd77380751e-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.752365 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c" (UID: "78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.752433 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a1d3e43-409c-48c7-a6ba-dcd77380751e-scripts" (OuterVolumeSpecName: "scripts") pod "6a1d3e43-409c-48c7-a6ba-dcd77380751e" (UID: "6a1d3e43-409c-48c7-a6ba-dcd77380751e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.752465 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6a1d3e43-409c-48c7-a6ba-dcd77380751e-var-run" (OuterVolumeSpecName: "var-run") pod "6a1d3e43-409c-48c7-a6ba-dcd77380751e" (UID: "6a1d3e43-409c-48c7-a6ba-dcd77380751e"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.752991 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a1d3e43-409c-48c7-a6ba-dcd77380751e-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "6a1d3e43-409c-48c7-a6ba-dcd77380751e" (UID: "6a1d3e43-409c-48c7-a6ba-dcd77380751e"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.753259 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c" (UID: "78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.765373 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-kube-api-access-q7vj2" (OuterVolumeSpecName: "kube-api-access-q7vj2") pod "78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c" (UID: "78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c"). InnerVolumeSpecName "kube-api-access-q7vj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.769791 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a1d3e43-409c-48c7-a6ba-dcd77380751e-kube-api-access-k25fr" (OuterVolumeSpecName: "kube-api-access-k25fr") pod "6a1d3e43-409c-48c7-a6ba-dcd77380751e" (UID: "6a1d3e43-409c-48c7-a6ba-dcd77380751e"). InnerVolumeSpecName "kube-api-access-k25fr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.770175 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c" (UID: "78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.774617 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-scripts" (OuterVolumeSpecName: "scripts") pod "78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c" (UID: "78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.780093 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c" (UID: "78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.780584 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c" (UID: "78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.854079 4998 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.854104 4998 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6a1d3e43-409c-48c7-a6ba-dcd77380751e-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.854117 4998 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6a1d3e43-409c-48c7-a6ba-dcd77380751e-var-run\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.854126 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.854134 4998 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.854142 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k25fr\" (UniqueName: \"kubernetes.io/projected/6a1d3e43-409c-48c7-a6ba-dcd77380751e-kube-api-access-k25fr\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.854152 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6a1d3e43-409c-48c7-a6ba-dcd77380751e-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.854160 4998 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.854170 4998 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.854179 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7vj2\" (UniqueName: \"kubernetes.io/projected/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-kube-api-access-q7vj2\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:08 crc kubenswrapper[4998]: I1203 16:23:08.854189 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:09 crc kubenswrapper[4998]: I1203 16:23:09.153418 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-xn4z8" Dec 03 16:23:09 crc kubenswrapper[4998]: I1203 16:23:09.153642 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-xn4z8" event={"ID":"78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c","Type":"ContainerDied","Data":"ba13d662cd465a86228e8355c796632f291f54b7fcfec3c5a6baa7bad8929396"} Dec 03 16:23:09 crc kubenswrapper[4998]: I1203 16:23:09.153905 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba13d662cd465a86228e8355c796632f291f54b7fcfec3c5a6baa7bad8929396" Dec 03 16:23:09 crc kubenswrapper[4998]: I1203 16:23:09.155243 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-7sxbw-config-l66rt" event={"ID":"6a1d3e43-409c-48c7-a6ba-dcd77380751e","Type":"ContainerDied","Data":"1a2dda1587ae7ba3c1adfeca96cdc488f3afc83114ef84a541db7d56717d1b49"} Dec 03 16:23:09 crc kubenswrapper[4998]: I1203 16:23:09.155284 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a2dda1587ae7ba3c1adfeca96cdc488f3afc83114ef84a541db7d56717d1b49" Dec 03 16:23:09 crc kubenswrapper[4998]: I1203 16:23:09.155326 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-7sxbw-config-l66rt" Dec 03 16:23:09 crc kubenswrapper[4998]: E1203 16:23:09.329583 4998 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78d5a8fb_51f1_464f_ae8a_f5ecd2be7a0c.slice\": RecentStats: unable to find data in memory cache]" Dec 03 16:23:09 crc kubenswrapper[4998]: I1203 16:23:09.659342 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-7sxbw-config-l66rt"] Dec 03 16:23:09 crc kubenswrapper[4998]: I1203 16:23:09.671603 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-7sxbw-config-l66rt"] Dec 03 16:23:09 crc kubenswrapper[4998]: I1203 16:23:09.687693 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a1d3e43-409c-48c7-a6ba-dcd77380751e" path="/var/lib/kubelet/pods/6a1d3e43-409c-48c7-a6ba-dcd77380751e/volumes" Dec 03 16:23:09 crc kubenswrapper[4998]: I1203 16:23:09.873534 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d01e9a78-9e2c-44c5-9602-4c8860123f88-etc-swift\") pod \"swift-storage-0\" (UID: \"d01e9a78-9e2c-44c5-9602-4c8860123f88\") " pod="openstack/swift-storage-0" Dec 03 16:23:09 crc kubenswrapper[4998]: I1203 16:23:09.880000 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d01e9a78-9e2c-44c5-9602-4c8860123f88-etc-swift\") pod \"swift-storage-0\" (UID: \"d01e9a78-9e2c-44c5-9602-4c8860123f88\") " pod="openstack/swift-storage-0" Dec 03 16:23:10 crc kubenswrapper[4998]: I1203 16:23:10.110362 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 16:23:10 crc kubenswrapper[4998]: I1203 16:23:10.424244 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="e0ef6669-5a8b-462d-9681-925859c86df1" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.106:5671: connect: connection refused" Dec 03 16:23:10 crc kubenswrapper[4998]: I1203 16:23:10.444774 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="aeb4a971-3c03-457c-b0df-1503701aac5b" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.105:5671: connect: connection refused" Dec 03 16:23:10 crc kubenswrapper[4998]: I1203 16:23:10.522252 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 03 16:23:10 crc kubenswrapper[4998]: W1203 16:23:10.528551 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd01e9a78_9e2c_44c5_9602_4c8860123f88.slice/crio-900265ce567f5a416d38f47e301bf1eb10ca48e3da5e431a86910cef3917600a WatchSource:0}: Error finding container 900265ce567f5a416d38f47e301bf1eb10ca48e3da5e431a86910cef3917600a: Status 404 returned error can't find the container with id 900265ce567f5a416d38f47e301bf1eb10ca48e3da5e431a86910cef3917600a Dec 03 16:23:10 crc kubenswrapper[4998]: I1203 16:23:10.835827 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-notifications-server-0" podUID="dbde2166-b545-4979-845a-8932845b9cd1" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.107:5671: connect: connection refused" Dec 03 16:23:11 crc kubenswrapper[4998]: I1203 16:23:11.181261 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d01e9a78-9e2c-44c5-9602-4c8860123f88","Type":"ContainerStarted","Data":"900265ce567f5a416d38f47e301bf1eb10ca48e3da5e431a86910cef3917600a"} Dec 03 16:23:12 crc kubenswrapper[4998]: I1203 16:23:12.192200 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d01e9a78-9e2c-44c5-9602-4c8860123f88","Type":"ContainerStarted","Data":"10622ec9cb97ad5e13db21d367c64ed878e4fcb326255ce4aa0730c4951b67ca"} Dec 03 16:23:12 crc kubenswrapper[4998]: I1203 16:23:12.192252 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d01e9a78-9e2c-44c5-9602-4c8860123f88","Type":"ContainerStarted","Data":"eeaeb48cefd940d7cc3502938d1c7b677bd5af13648facda75f1a92fce603d02"} Dec 03 16:23:12 crc kubenswrapper[4998]: I1203 16:23:12.192265 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d01e9a78-9e2c-44c5-9602-4c8860123f88","Type":"ContainerStarted","Data":"637eeb008b504aeb152d16c764eb2840a1e7769eb1919b598ecb8171823d7953"} Dec 03 16:23:12 crc kubenswrapper[4998]: I1203 16:23:12.192276 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d01e9a78-9e2c-44c5-9602-4c8860123f88","Type":"ContainerStarted","Data":"e8562cd2c8995ba3e93e8b6d686bbac55d6067ae8653eb9925734450da97e228"} Dec 03 16:23:13 crc kubenswrapper[4998]: I1203 16:23:13.150818 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:13 crc kubenswrapper[4998]: I1203 16:23:13.154375 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:13 crc kubenswrapper[4998]: I1203 16:23:13.215060 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d01e9a78-9e2c-44c5-9602-4c8860123f88","Type":"ContainerStarted","Data":"9e71f7474d05629dbc07585767980b48963963fd04d45ce2b779cd7f919d321c"} Dec 03 16:23:13 crc kubenswrapper[4998]: I1203 16:23:13.215114 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d01e9a78-9e2c-44c5-9602-4c8860123f88","Type":"ContainerStarted","Data":"b1701e3fd2c8fb5de80bc740769fdf9e0cba388b6690a0735484793376e8cd5c"} Dec 03 16:23:13 crc kubenswrapper[4998]: I1203 16:23:13.219271 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:14 crc kubenswrapper[4998]: I1203 16:23:14.225506 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d01e9a78-9e2c-44c5-9602-4c8860123f88","Type":"ContainerStarted","Data":"f8ac7b0952b839f3b8774366c9048e675c08c9b609e6ea3dc67beb0bdba98c53"} Dec 03 16:23:14 crc kubenswrapper[4998]: I1203 16:23:14.225896 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d01e9a78-9e2c-44c5-9602-4c8860123f88","Type":"ContainerStarted","Data":"f40712dee3faae55bd174c4d87a766c7f77201d1054344ad856abc1a7ee8d053"} Dec 03 16:23:15 crc kubenswrapper[4998]: I1203 16:23:15.024288 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 16:23:15 crc kubenswrapper[4998]: I1203 16:23:15.238063 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d01e9a78-9e2c-44c5-9602-4c8860123f88","Type":"ContainerStarted","Data":"d76db74f107321082a20d14ef26871ee67169dffe624410d5839efce5dc232af"} Dec 03 16:23:15 crc kubenswrapper[4998]: I1203 16:23:15.238105 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d01e9a78-9e2c-44c5-9602-4c8860123f88","Type":"ContainerStarted","Data":"44810949fa59366384f7bf9ecabaa3e1ffc1f6177d1e33bc59d393915972918e"} Dec 03 16:23:15 crc kubenswrapper[4998]: I1203 16:23:15.238116 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d01e9a78-9e2c-44c5-9602-4c8860123f88","Type":"ContainerStarted","Data":"bcbe279ef74c7d5f7a72140431b1ee37810f48ff9f3fa4efb8f73cb9fea12e48"} Dec 03 16:23:15 crc kubenswrapper[4998]: I1203 16:23:15.238245 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="92d2d9f9-b0c3-40f8-981a-41ec47f030e1" containerName="prometheus" containerID="cri-o://39b4425a9a234b6550ab87b9df86280de262b2abf3cd4770ea0623e4f4b32977" gracePeriod=600 Dec 03 16:23:15 crc kubenswrapper[4998]: I1203 16:23:15.238284 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="92d2d9f9-b0c3-40f8-981a-41ec47f030e1" containerName="thanos-sidecar" containerID="cri-o://99c43f95cd03544481cc3d9f5eefe61127fd464b7135edd8f2e2587a46506ef9" gracePeriod=600 Dec 03 16:23:15 crc kubenswrapper[4998]: I1203 16:23:15.238305 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="92d2d9f9-b0c3-40f8-981a-41ec47f030e1" containerName="config-reloader" containerID="cri-o://392644aa28c6eff25fa63b88178eef2345ac2185aecdb1961b66fdaee4d65dcc" gracePeriod=600 Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.103419 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.178427 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rz4xj\" (UniqueName: \"kubernetes.io/projected/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-kube-api-access-rz4xj\") pod \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.178520 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-config-out\") pod \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.178551 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-thanos-prometheus-http-client-file\") pod \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.178704 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c9889e3e-2478-469f-b92d-e203678c4927\") pod \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.178728 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-prometheus-metric-storage-rulefiles-0\") pod \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.178779 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-tls-assets\") pod \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.178891 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-web-config\") pod \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.178920 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-config\") pod \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\" (UID: \"92d2d9f9-b0c3-40f8-981a-41ec47f030e1\") " Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.180732 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "92d2d9f9-b0c3-40f8-981a-41ec47f030e1" (UID: "92d2d9f9-b0c3-40f8-981a-41ec47f030e1"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.187635 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-kube-api-access-rz4xj" (OuterVolumeSpecName: "kube-api-access-rz4xj") pod "92d2d9f9-b0c3-40f8-981a-41ec47f030e1" (UID: "92d2d9f9-b0c3-40f8-981a-41ec47f030e1"). InnerVolumeSpecName "kube-api-access-rz4xj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.188010 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "92d2d9f9-b0c3-40f8-981a-41ec47f030e1" (UID: "92d2d9f9-b0c3-40f8-981a-41ec47f030e1"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.189095 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-config-out" (OuterVolumeSpecName: "config-out") pod "92d2d9f9-b0c3-40f8-981a-41ec47f030e1" (UID: "92d2d9f9-b0c3-40f8-981a-41ec47f030e1"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.191308 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-config" (OuterVolumeSpecName: "config") pod "92d2d9f9-b0c3-40f8-981a-41ec47f030e1" (UID: "92d2d9f9-b0c3-40f8-981a-41ec47f030e1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.191948 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "92d2d9f9-b0c3-40f8-981a-41ec47f030e1" (UID: "92d2d9f9-b0c3-40f8-981a-41ec47f030e1"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.210477 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-web-config" (OuterVolumeSpecName: "web-config") pod "92d2d9f9-b0c3-40f8-981a-41ec47f030e1" (UID: "92d2d9f9-b0c3-40f8-981a-41ec47f030e1"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.211243 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c9889e3e-2478-469f-b92d-e203678c4927" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "92d2d9f9-b0c3-40f8-981a-41ec47f030e1" (UID: "92d2d9f9-b0c3-40f8-981a-41ec47f030e1"). InnerVolumeSpecName "pvc-c9889e3e-2478-469f-b92d-e203678c4927". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.249690 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d01e9a78-9e2c-44c5-9602-4c8860123f88","Type":"ContainerStarted","Data":"63bc6d37c98647a3c2928d89fa1ca11e01b1d5a165981ecd53480065b7cf86af"} Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.250914 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d01e9a78-9e2c-44c5-9602-4c8860123f88","Type":"ContainerStarted","Data":"e27b34ecb873d0ee1c8d70d207a9f301266fbc582e5a3be694cb41e9125876fb"} Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.251035 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d01e9a78-9e2c-44c5-9602-4c8860123f88","Type":"ContainerStarted","Data":"c55f4c5470ea5973a1e642438a15d2635a1b8d7319683c10dfd8c6977305b7de"} Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.251169 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d01e9a78-9e2c-44c5-9602-4c8860123f88","Type":"ContainerStarted","Data":"4048c2e8c1c4780a89ba2b767b5108db9c8030f3e8d1eada23f8c95b0d62cb17"} Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.253197 4998 generic.go:334] "Generic (PLEG): container finished" podID="92d2d9f9-b0c3-40f8-981a-41ec47f030e1" containerID="99c43f95cd03544481cc3d9f5eefe61127fd464b7135edd8f2e2587a46506ef9" exitCode=0 Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.253359 4998 generic.go:334] "Generic (PLEG): container finished" podID="92d2d9f9-b0c3-40f8-981a-41ec47f030e1" containerID="392644aa28c6eff25fa63b88178eef2345ac2185aecdb1961b66fdaee4d65dcc" exitCode=0 Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.253428 4998 generic.go:334] "Generic (PLEG): container finished" podID="92d2d9f9-b0c3-40f8-981a-41ec47f030e1" containerID="39b4425a9a234b6550ab87b9df86280de262b2abf3cd4770ea0623e4f4b32977" exitCode=0 Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.253400 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.253358 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92d2d9f9-b0c3-40f8-981a-41ec47f030e1","Type":"ContainerDied","Data":"99c43f95cd03544481cc3d9f5eefe61127fd464b7135edd8f2e2587a46506ef9"} Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.253663 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92d2d9f9-b0c3-40f8-981a-41ec47f030e1","Type":"ContainerDied","Data":"392644aa28c6eff25fa63b88178eef2345ac2185aecdb1961b66fdaee4d65dcc"} Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.253676 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92d2d9f9-b0c3-40f8-981a-41ec47f030e1","Type":"ContainerDied","Data":"39b4425a9a234b6550ab87b9df86280de262b2abf3cd4770ea0623e4f4b32977"} Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.253685 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92d2d9f9-b0c3-40f8-981a-41ec47f030e1","Type":"ContainerDied","Data":"85e455c8ca11f61be037396af6a3f899fc6743b182bf018a7a79b9b2f7d475e2"} Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.253692 4998 scope.go:117] "RemoveContainer" containerID="99c43f95cd03544481cc3d9f5eefe61127fd464b7135edd8f2e2587a46506ef9" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.273023 4998 scope.go:117] "RemoveContainer" containerID="392644aa28c6eff25fa63b88178eef2345ac2185aecdb1961b66fdaee4d65dcc" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.283632 4998 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-web-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.283665 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.283675 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rz4xj\" (UniqueName: \"kubernetes.io/projected/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-kube-api-access-rz4xj\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.283687 4998 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-config-out\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.283696 4998 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.283705 4998 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.283734 4998 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-c9889e3e-2478-469f-b92d-e203678c4927\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c9889e3e-2478-469f-b92d-e203678c4927\") on node \"crc\" " Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.283745 4998 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/92d2d9f9-b0c3-40f8-981a-41ec47f030e1-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.291712 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.641533742 podStartE2EDuration="40.291692911s" podCreationTimestamp="2025-12-03 16:22:36 +0000 UTC" firstStartedPulling="2025-12-03 16:23:10.531093774 +0000 UTC m=+1169.142793997" lastFinishedPulling="2025-12-03 16:23:14.181252943 +0000 UTC m=+1172.792953166" observedRunningTime="2025-12-03 16:23:16.287592848 +0000 UTC m=+1174.899293071" watchObservedRunningTime="2025-12-03 16:23:16.291692911 +0000 UTC m=+1174.903393134" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.299037 4998 scope.go:117] "RemoveContainer" containerID="39b4425a9a234b6550ab87b9df86280de262b2abf3cd4770ea0623e4f4b32977" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.328414 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.330565 4998 scope.go:117] "RemoveContainer" containerID="174a248e64d7d0441f1074334990836bdc3697b8f54e27015258af83a045fcd0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.342791 4998 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.342980 4998 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-c9889e3e-2478-469f-b92d-e203678c4927" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c9889e3e-2478-469f-b92d-e203678c4927") on node "crc" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.354059 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.361056 4998 scope.go:117] "RemoveContainer" containerID="99c43f95cd03544481cc3d9f5eefe61127fd464b7135edd8f2e2587a46506ef9" Dec 03 16:23:16 crc kubenswrapper[4998]: E1203 16:23:16.363182 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99c43f95cd03544481cc3d9f5eefe61127fd464b7135edd8f2e2587a46506ef9\": container with ID starting with 99c43f95cd03544481cc3d9f5eefe61127fd464b7135edd8f2e2587a46506ef9 not found: ID does not exist" containerID="99c43f95cd03544481cc3d9f5eefe61127fd464b7135edd8f2e2587a46506ef9" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.363233 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99c43f95cd03544481cc3d9f5eefe61127fd464b7135edd8f2e2587a46506ef9"} err="failed to get container status \"99c43f95cd03544481cc3d9f5eefe61127fd464b7135edd8f2e2587a46506ef9\": rpc error: code = NotFound desc = could not find container \"99c43f95cd03544481cc3d9f5eefe61127fd464b7135edd8f2e2587a46506ef9\": container with ID starting with 99c43f95cd03544481cc3d9f5eefe61127fd464b7135edd8f2e2587a46506ef9 not found: ID does not exist" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.363263 4998 scope.go:117] "RemoveContainer" containerID="392644aa28c6eff25fa63b88178eef2345ac2185aecdb1961b66fdaee4d65dcc" Dec 03 16:23:16 crc kubenswrapper[4998]: E1203 16:23:16.372908 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"392644aa28c6eff25fa63b88178eef2345ac2185aecdb1961b66fdaee4d65dcc\": container with ID starting with 392644aa28c6eff25fa63b88178eef2345ac2185aecdb1961b66fdaee4d65dcc not found: ID does not exist" containerID="392644aa28c6eff25fa63b88178eef2345ac2185aecdb1961b66fdaee4d65dcc" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.372955 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"392644aa28c6eff25fa63b88178eef2345ac2185aecdb1961b66fdaee4d65dcc"} err="failed to get container status \"392644aa28c6eff25fa63b88178eef2345ac2185aecdb1961b66fdaee4d65dcc\": rpc error: code = NotFound desc = could not find container \"392644aa28c6eff25fa63b88178eef2345ac2185aecdb1961b66fdaee4d65dcc\": container with ID starting with 392644aa28c6eff25fa63b88178eef2345ac2185aecdb1961b66fdaee4d65dcc not found: ID does not exist" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.372979 4998 scope.go:117] "RemoveContainer" containerID="39b4425a9a234b6550ab87b9df86280de262b2abf3cd4770ea0623e4f4b32977" Dec 03 16:23:16 crc kubenswrapper[4998]: E1203 16:23:16.374307 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39b4425a9a234b6550ab87b9df86280de262b2abf3cd4770ea0623e4f4b32977\": container with ID starting with 39b4425a9a234b6550ab87b9df86280de262b2abf3cd4770ea0623e4f4b32977 not found: ID does not exist" containerID="39b4425a9a234b6550ab87b9df86280de262b2abf3cd4770ea0623e4f4b32977" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.374336 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39b4425a9a234b6550ab87b9df86280de262b2abf3cd4770ea0623e4f4b32977"} err="failed to get container status \"39b4425a9a234b6550ab87b9df86280de262b2abf3cd4770ea0623e4f4b32977\": rpc error: code = NotFound desc = could not find container \"39b4425a9a234b6550ab87b9df86280de262b2abf3cd4770ea0623e4f4b32977\": container with ID starting with 39b4425a9a234b6550ab87b9df86280de262b2abf3cd4770ea0623e4f4b32977 not found: ID does not exist" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.374350 4998 scope.go:117] "RemoveContainer" containerID="174a248e64d7d0441f1074334990836bdc3697b8f54e27015258af83a045fcd0" Dec 03 16:23:16 crc kubenswrapper[4998]: E1203 16:23:16.374771 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"174a248e64d7d0441f1074334990836bdc3697b8f54e27015258af83a045fcd0\": container with ID starting with 174a248e64d7d0441f1074334990836bdc3697b8f54e27015258af83a045fcd0 not found: ID does not exist" containerID="174a248e64d7d0441f1074334990836bdc3697b8f54e27015258af83a045fcd0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.374825 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"174a248e64d7d0441f1074334990836bdc3697b8f54e27015258af83a045fcd0"} err="failed to get container status \"174a248e64d7d0441f1074334990836bdc3697b8f54e27015258af83a045fcd0\": rpc error: code = NotFound desc = could not find container \"174a248e64d7d0441f1074334990836bdc3697b8f54e27015258af83a045fcd0\": container with ID starting with 174a248e64d7d0441f1074334990836bdc3697b8f54e27015258af83a045fcd0 not found: ID does not exist" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.374855 4998 scope.go:117] "RemoveContainer" containerID="99c43f95cd03544481cc3d9f5eefe61127fd464b7135edd8f2e2587a46506ef9" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.376243 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99c43f95cd03544481cc3d9f5eefe61127fd464b7135edd8f2e2587a46506ef9"} err="failed to get container status \"99c43f95cd03544481cc3d9f5eefe61127fd464b7135edd8f2e2587a46506ef9\": rpc error: code = NotFound desc = could not find container \"99c43f95cd03544481cc3d9f5eefe61127fd464b7135edd8f2e2587a46506ef9\": container with ID starting with 99c43f95cd03544481cc3d9f5eefe61127fd464b7135edd8f2e2587a46506ef9 not found: ID does not exist" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.376268 4998 scope.go:117] "RemoveContainer" containerID="392644aa28c6eff25fa63b88178eef2345ac2185aecdb1961b66fdaee4d65dcc" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.376581 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"392644aa28c6eff25fa63b88178eef2345ac2185aecdb1961b66fdaee4d65dcc"} err="failed to get container status \"392644aa28c6eff25fa63b88178eef2345ac2185aecdb1961b66fdaee4d65dcc\": rpc error: code = NotFound desc = could not find container \"392644aa28c6eff25fa63b88178eef2345ac2185aecdb1961b66fdaee4d65dcc\": container with ID starting with 392644aa28c6eff25fa63b88178eef2345ac2185aecdb1961b66fdaee4d65dcc not found: ID does not exist" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.376601 4998 scope.go:117] "RemoveContainer" containerID="39b4425a9a234b6550ab87b9df86280de262b2abf3cd4770ea0623e4f4b32977" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.376818 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39b4425a9a234b6550ab87b9df86280de262b2abf3cd4770ea0623e4f4b32977"} err="failed to get container status \"39b4425a9a234b6550ab87b9df86280de262b2abf3cd4770ea0623e4f4b32977\": rpc error: code = NotFound desc = could not find container \"39b4425a9a234b6550ab87b9df86280de262b2abf3cd4770ea0623e4f4b32977\": container with ID starting with 39b4425a9a234b6550ab87b9df86280de262b2abf3cd4770ea0623e4f4b32977 not found: ID does not exist" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.376837 4998 scope.go:117] "RemoveContainer" containerID="174a248e64d7d0441f1074334990836bdc3697b8f54e27015258af83a045fcd0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.378298 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"174a248e64d7d0441f1074334990836bdc3697b8f54e27015258af83a045fcd0"} err="failed to get container status \"174a248e64d7d0441f1074334990836bdc3697b8f54e27015258af83a045fcd0\": rpc error: code = NotFound desc = could not find container \"174a248e64d7d0441f1074334990836bdc3697b8f54e27015258af83a045fcd0\": container with ID starting with 174a248e64d7d0441f1074334990836bdc3697b8f54e27015258af83a045fcd0 not found: ID does not exist" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.378350 4998 scope.go:117] "RemoveContainer" containerID="99c43f95cd03544481cc3d9f5eefe61127fd464b7135edd8f2e2587a46506ef9" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.378902 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99c43f95cd03544481cc3d9f5eefe61127fd464b7135edd8f2e2587a46506ef9"} err="failed to get container status \"99c43f95cd03544481cc3d9f5eefe61127fd464b7135edd8f2e2587a46506ef9\": rpc error: code = NotFound desc = could not find container \"99c43f95cd03544481cc3d9f5eefe61127fd464b7135edd8f2e2587a46506ef9\": container with ID starting with 99c43f95cd03544481cc3d9f5eefe61127fd464b7135edd8f2e2587a46506ef9 not found: ID does not exist" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.378961 4998 scope.go:117] "RemoveContainer" containerID="392644aa28c6eff25fa63b88178eef2345ac2185aecdb1961b66fdaee4d65dcc" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.382807 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 16:23:16 crc kubenswrapper[4998]: E1203 16:23:16.383170 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c" containerName="swift-ring-rebalance" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.383184 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c" containerName="swift-ring-rebalance" Dec 03 16:23:16 crc kubenswrapper[4998]: E1203 16:23:16.383200 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a1d3e43-409c-48c7-a6ba-dcd77380751e" containerName="ovn-config" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.383205 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a1d3e43-409c-48c7-a6ba-dcd77380751e" containerName="ovn-config" Dec 03 16:23:16 crc kubenswrapper[4998]: E1203 16:23:16.383221 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d2d9f9-b0c3-40f8-981a-41ec47f030e1" containerName="prometheus" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.383228 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d2d9f9-b0c3-40f8-981a-41ec47f030e1" containerName="prometheus" Dec 03 16:23:16 crc kubenswrapper[4998]: E1203 16:23:16.383238 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d2d9f9-b0c3-40f8-981a-41ec47f030e1" containerName="thanos-sidecar" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.383244 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d2d9f9-b0c3-40f8-981a-41ec47f030e1" containerName="thanos-sidecar" Dec 03 16:23:16 crc kubenswrapper[4998]: E1203 16:23:16.383259 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d2d9f9-b0c3-40f8-981a-41ec47f030e1" containerName="config-reloader" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.383264 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d2d9f9-b0c3-40f8-981a-41ec47f030e1" containerName="config-reloader" Dec 03 16:23:16 crc kubenswrapper[4998]: E1203 16:23:16.383277 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d2d9f9-b0c3-40f8-981a-41ec47f030e1" containerName="init-config-reloader" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.383282 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d2d9f9-b0c3-40f8-981a-41ec47f030e1" containerName="init-config-reloader" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.383433 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a1d3e43-409c-48c7-a6ba-dcd77380751e" containerName="ovn-config" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.383445 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c" containerName="swift-ring-rebalance" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.383456 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d2d9f9-b0c3-40f8-981a-41ec47f030e1" containerName="thanos-sidecar" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.383465 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d2d9f9-b0c3-40f8-981a-41ec47f030e1" containerName="config-reloader" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.383481 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d2d9f9-b0c3-40f8-981a-41ec47f030e1" containerName="prometheus" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.384910 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.386103 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"392644aa28c6eff25fa63b88178eef2345ac2185aecdb1961b66fdaee4d65dcc"} err="failed to get container status \"392644aa28c6eff25fa63b88178eef2345ac2185aecdb1961b66fdaee4d65dcc\": rpc error: code = NotFound desc = could not find container \"392644aa28c6eff25fa63b88178eef2345ac2185aecdb1961b66fdaee4d65dcc\": container with ID starting with 392644aa28c6eff25fa63b88178eef2345ac2185aecdb1961b66fdaee4d65dcc not found: ID does not exist" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.386131 4998 scope.go:117] "RemoveContainer" containerID="39b4425a9a234b6550ab87b9df86280de262b2abf3cd4770ea0623e4f4b32977" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.387149 4998 reconciler_common.go:293] "Volume detached for volume \"pvc-c9889e3e-2478-469f-b92d-e203678c4927\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c9889e3e-2478-469f-b92d-e203678c4927\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.388381 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-fx7rj" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.388524 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.388648 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.388830 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.388967 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.389075 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.395310 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.395792 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39b4425a9a234b6550ab87b9df86280de262b2abf3cd4770ea0623e4f4b32977"} err="failed to get container status \"39b4425a9a234b6550ab87b9df86280de262b2abf3cd4770ea0623e4f4b32977\": rpc error: code = NotFound desc = could not find container \"39b4425a9a234b6550ab87b9df86280de262b2abf3cd4770ea0623e4f4b32977\": container with ID starting with 39b4425a9a234b6550ab87b9df86280de262b2abf3cd4770ea0623e4f4b32977 not found: ID does not exist" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.395840 4998 scope.go:117] "RemoveContainer" containerID="174a248e64d7d0441f1074334990836bdc3697b8f54e27015258af83a045fcd0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.397201 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"174a248e64d7d0441f1074334990836bdc3697b8f54e27015258af83a045fcd0"} err="failed to get container status \"174a248e64d7d0441f1074334990836bdc3697b8f54e27015258af83a045fcd0\": rpc error: code = NotFound desc = could not find container \"174a248e64d7d0441f1074334990836bdc3697b8f54e27015258af83a045fcd0\": container with ID starting with 174a248e64d7d0441f1074334990836bdc3697b8f54e27015258af83a045fcd0 not found: ID does not exist" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.399488 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.489042 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a70537b1-93bc-48c3-9bef-fd9030158cad-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.489147 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.489214 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-config\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.489237 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a70537b1-93bc-48c3-9bef-fd9030158cad-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.489304 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.489371 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bh8n\" (UniqueName: \"kubernetes.io/projected/a70537b1-93bc-48c3-9bef-fd9030158cad-kube-api-access-6bh8n\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.489411 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.489437 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/a70537b1-93bc-48c3-9bef-fd9030158cad-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.489478 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c9889e3e-2478-469f-b92d-e203678c4927\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c9889e3e-2478-469f-b92d-e203678c4927\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.489526 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.489574 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.584526 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-946dbfbcf-7n244"] Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.586195 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-946dbfbcf-7n244" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.592428 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.593367 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.593397 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/a70537b1-93bc-48c3-9bef-fd9030158cad-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.593424 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c9889e3e-2478-469f-b92d-e203678c4927\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c9889e3e-2478-469f-b92d-e203678c4927\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.593453 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.593487 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.593514 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a70537b1-93bc-48c3-9bef-fd9030158cad-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.593563 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.593591 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-config\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.593605 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a70537b1-93bc-48c3-9bef-fd9030158cad-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.593631 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.593656 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bh8n\" (UniqueName: \"kubernetes.io/projected/a70537b1-93bc-48c3-9bef-fd9030158cad-kube-api-access-6bh8n\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.597655 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.598202 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/a70537b1-93bc-48c3-9bef-fd9030158cad-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.599385 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-946dbfbcf-7n244"] Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.601209 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.604093 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a70537b1-93bc-48c3-9bef-fd9030158cad-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.606191 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-config\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.606726 4998 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.606773 4998 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c9889e3e-2478-469f-b92d-e203678c4927\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c9889e3e-2478-469f-b92d-e203678c4927\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/afadf41a682809fbeb8f0a8dc0e25f598b7474e7bc0a2ee0471a59f3f2d1c942/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.607489 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.609150 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.609173 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.609729 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a70537b1-93bc-48c3-9bef-fd9030158cad-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.629099 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bh8n\" (UniqueName: \"kubernetes.io/projected/a70537b1-93bc-48c3-9bef-fd9030158cad-kube-api-access-6bh8n\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.658079 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c9889e3e-2478-469f-b92d-e203678c4927\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c9889e3e-2478-469f-b92d-e203678c4927\") pod \"prometheus-metric-storage-0\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.695071 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-dns-swift-storage-0\") pod \"dnsmasq-dns-946dbfbcf-7n244\" (UID: \"b5c15d74-1721-40f9-bb22-162a95fd98c1\") " pod="openstack/dnsmasq-dns-946dbfbcf-7n244" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.695131 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-config\") pod \"dnsmasq-dns-946dbfbcf-7n244\" (UID: \"b5c15d74-1721-40f9-bb22-162a95fd98c1\") " pod="openstack/dnsmasq-dns-946dbfbcf-7n244" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.695253 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-ovsdbserver-sb\") pod \"dnsmasq-dns-946dbfbcf-7n244\" (UID: \"b5c15d74-1721-40f9-bb22-162a95fd98c1\") " pod="openstack/dnsmasq-dns-946dbfbcf-7n244" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.695333 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-ovsdbserver-nb\") pod \"dnsmasq-dns-946dbfbcf-7n244\" (UID: \"b5c15d74-1721-40f9-bb22-162a95fd98c1\") " pod="openstack/dnsmasq-dns-946dbfbcf-7n244" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.695357 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-dns-svc\") pod \"dnsmasq-dns-946dbfbcf-7n244\" (UID: \"b5c15d74-1721-40f9-bb22-162a95fd98c1\") " pod="openstack/dnsmasq-dns-946dbfbcf-7n244" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.695532 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vlnw\" (UniqueName: \"kubernetes.io/projected/b5c15d74-1721-40f9-bb22-162a95fd98c1-kube-api-access-8vlnw\") pod \"dnsmasq-dns-946dbfbcf-7n244\" (UID: \"b5c15d74-1721-40f9-bb22-162a95fd98c1\") " pod="openstack/dnsmasq-dns-946dbfbcf-7n244" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.736796 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.797276 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-config\") pod \"dnsmasq-dns-946dbfbcf-7n244\" (UID: \"b5c15d74-1721-40f9-bb22-162a95fd98c1\") " pod="openstack/dnsmasq-dns-946dbfbcf-7n244" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.797353 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-ovsdbserver-sb\") pod \"dnsmasq-dns-946dbfbcf-7n244\" (UID: \"b5c15d74-1721-40f9-bb22-162a95fd98c1\") " pod="openstack/dnsmasq-dns-946dbfbcf-7n244" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.797390 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-ovsdbserver-nb\") pod \"dnsmasq-dns-946dbfbcf-7n244\" (UID: \"b5c15d74-1721-40f9-bb22-162a95fd98c1\") " pod="openstack/dnsmasq-dns-946dbfbcf-7n244" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.797409 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-dns-svc\") pod \"dnsmasq-dns-946dbfbcf-7n244\" (UID: \"b5c15d74-1721-40f9-bb22-162a95fd98c1\") " pod="openstack/dnsmasq-dns-946dbfbcf-7n244" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.797433 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vlnw\" (UniqueName: \"kubernetes.io/projected/b5c15d74-1721-40f9-bb22-162a95fd98c1-kube-api-access-8vlnw\") pod \"dnsmasq-dns-946dbfbcf-7n244\" (UID: \"b5c15d74-1721-40f9-bb22-162a95fd98c1\") " pod="openstack/dnsmasq-dns-946dbfbcf-7n244" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.797485 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-dns-swift-storage-0\") pod \"dnsmasq-dns-946dbfbcf-7n244\" (UID: \"b5c15d74-1721-40f9-bb22-162a95fd98c1\") " pod="openstack/dnsmasq-dns-946dbfbcf-7n244" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.798145 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-config\") pod \"dnsmasq-dns-946dbfbcf-7n244\" (UID: \"b5c15d74-1721-40f9-bb22-162a95fd98c1\") " pod="openstack/dnsmasq-dns-946dbfbcf-7n244" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.798187 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-dns-swift-storage-0\") pod \"dnsmasq-dns-946dbfbcf-7n244\" (UID: \"b5c15d74-1721-40f9-bb22-162a95fd98c1\") " pod="openstack/dnsmasq-dns-946dbfbcf-7n244" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.798679 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-ovsdbserver-nb\") pod \"dnsmasq-dns-946dbfbcf-7n244\" (UID: \"b5c15d74-1721-40f9-bb22-162a95fd98c1\") " pod="openstack/dnsmasq-dns-946dbfbcf-7n244" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.803788 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-dns-svc\") pod \"dnsmasq-dns-946dbfbcf-7n244\" (UID: \"b5c15d74-1721-40f9-bb22-162a95fd98c1\") " pod="openstack/dnsmasq-dns-946dbfbcf-7n244" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.805821 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-ovsdbserver-sb\") pod \"dnsmasq-dns-946dbfbcf-7n244\" (UID: \"b5c15d74-1721-40f9-bb22-162a95fd98c1\") " pod="openstack/dnsmasq-dns-946dbfbcf-7n244" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.819180 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vlnw\" (UniqueName: \"kubernetes.io/projected/b5c15d74-1721-40f9-bb22-162a95fd98c1-kube-api-access-8vlnw\") pod \"dnsmasq-dns-946dbfbcf-7n244\" (UID: \"b5c15d74-1721-40f9-bb22-162a95fd98c1\") " pod="openstack/dnsmasq-dns-946dbfbcf-7n244" Dec 03 16:23:16 crc kubenswrapper[4998]: I1203 16:23:16.982288 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-946dbfbcf-7n244" Dec 03 16:23:17 crc kubenswrapper[4998]: I1203 16:23:17.160600 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 16:23:17 crc kubenswrapper[4998]: W1203 16:23:17.165348 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda70537b1_93bc_48c3_9bef_fd9030158cad.slice/crio-42cc47506f3d7b982ad83a6d55f67124b3d014b607d4dc73e61b9a42e1f91d65 WatchSource:0}: Error finding container 42cc47506f3d7b982ad83a6d55f67124b3d014b607d4dc73e61b9a42e1f91d65: Status 404 returned error can't find the container with id 42cc47506f3d7b982ad83a6d55f67124b3d014b607d4dc73e61b9a42e1f91d65 Dec 03 16:23:17 crc kubenswrapper[4998]: I1203 16:23:17.263266 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a70537b1-93bc-48c3-9bef-fd9030158cad","Type":"ContainerStarted","Data":"42cc47506f3d7b982ad83a6d55f67124b3d014b607d4dc73e61b9a42e1f91d65"} Dec 03 16:23:17 crc kubenswrapper[4998]: I1203 16:23:17.418738 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-946dbfbcf-7n244"] Dec 03 16:23:17 crc kubenswrapper[4998]: W1203 16:23:17.428836 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5c15d74_1721_40f9_bb22_162a95fd98c1.slice/crio-5d41092474acf4cc286d3ab50aeae4bac75de306e3f04da4736a2d5a7fd2540c WatchSource:0}: Error finding container 5d41092474acf4cc286d3ab50aeae4bac75de306e3f04da4736a2d5a7fd2540c: Status 404 returned error can't find the container with id 5d41092474acf4cc286d3ab50aeae4bac75de306e3f04da4736a2d5a7fd2540c Dec 03 16:23:17 crc kubenswrapper[4998]: I1203 16:23:17.687653 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92d2d9f9-b0c3-40f8-981a-41ec47f030e1" path="/var/lib/kubelet/pods/92d2d9f9-b0c3-40f8-981a-41ec47f030e1/volumes" Dec 03 16:23:18 crc kubenswrapper[4998]: I1203 16:23:18.276804 4998 generic.go:334] "Generic (PLEG): container finished" podID="b5c15d74-1721-40f9-bb22-162a95fd98c1" containerID="c3e6d8b3107f8f81a007fb86a9f63de459dd96469026a585ee40be984abc2b64" exitCode=0 Dec 03 16:23:18 crc kubenswrapper[4998]: I1203 16:23:18.276868 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-946dbfbcf-7n244" event={"ID":"b5c15d74-1721-40f9-bb22-162a95fd98c1","Type":"ContainerDied","Data":"c3e6d8b3107f8f81a007fb86a9f63de459dd96469026a585ee40be984abc2b64"} Dec 03 16:23:18 crc kubenswrapper[4998]: I1203 16:23:18.276904 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-946dbfbcf-7n244" event={"ID":"b5c15d74-1721-40f9-bb22-162a95fd98c1","Type":"ContainerStarted","Data":"5d41092474acf4cc286d3ab50aeae4bac75de306e3f04da4736a2d5a7fd2540c"} Dec 03 16:23:19 crc kubenswrapper[4998]: I1203 16:23:19.293153 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-946dbfbcf-7n244" event={"ID":"b5c15d74-1721-40f9-bb22-162a95fd98c1","Type":"ContainerStarted","Data":"ed81737b4f2a66d75bfdb335123457226be42a4d0163ce2162b676f760450e00"} Dec 03 16:23:19 crc kubenswrapper[4998]: I1203 16:23:19.293576 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-946dbfbcf-7n244" Dec 03 16:23:19 crc kubenswrapper[4998]: I1203 16:23:19.331193 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-946dbfbcf-7n244" podStartSLOduration=3.331166395 podStartE2EDuration="3.331166395s" podCreationTimestamp="2025-12-03 16:23:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:23:19.319176423 +0000 UTC m=+1177.930876676" watchObservedRunningTime="2025-12-03 16:23:19.331166395 +0000 UTC m=+1177.942866648" Dec 03 16:23:20 crc kubenswrapper[4998]: I1203 16:23:20.423132 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:23:20 crc kubenswrapper[4998]: I1203 16:23:20.442054 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 16:23:20 crc kubenswrapper[4998]: I1203 16:23:20.834912 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-notifications-server-0" Dec 03 16:23:21 crc kubenswrapper[4998]: I1203 16:23:21.315981 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a70537b1-93bc-48c3-9bef-fd9030158cad","Type":"ContainerStarted","Data":"9bd7f454c517ad30a0d2fb22370f32404af1a46c10a6bbd97476e74751b2c2c1"} Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.099732 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-rsc5q"] Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.101911 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rsc5q" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.117708 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-rsc5q"] Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.178287 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-b4hsk"] Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.179781 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-b4hsk" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.210347 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20531474-cb0d-4ce8-90fc-e9974acc6d1a-operator-scripts\") pod \"cinder-db-create-b4hsk\" (UID: \"20531474-cb0d-4ce8-90fc-e9974acc6d1a\") " pod="openstack/cinder-db-create-b4hsk" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.210407 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a4bd716-a01d-4e69-90ee-ae27b0e77c85-operator-scripts\") pod \"barbican-db-create-rsc5q\" (UID: \"2a4bd716-a01d-4e69-90ee-ae27b0e77c85\") " pod="openstack/barbican-db-create-rsc5q" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.210504 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfbld\" (UniqueName: \"kubernetes.io/projected/2a4bd716-a01d-4e69-90ee-ae27b0e77c85-kube-api-access-qfbld\") pod \"barbican-db-create-rsc5q\" (UID: \"2a4bd716-a01d-4e69-90ee-ae27b0e77c85\") " pod="openstack/barbican-db-create-rsc5q" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.210545 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf7t9\" (UniqueName: \"kubernetes.io/projected/20531474-cb0d-4ce8-90fc-e9974acc6d1a-kube-api-access-wf7t9\") pod \"cinder-db-create-b4hsk\" (UID: \"20531474-cb0d-4ce8-90fc-e9974acc6d1a\") " pod="openstack/cinder-db-create-b4hsk" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.240908 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-b4hsk"] Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.291014 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-7a2f-account-create-update-fjb4t"] Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.292004 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7a2f-account-create-update-fjb4t" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.305931 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.311898 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxc2h\" (UniqueName: \"kubernetes.io/projected/fecdd245-3377-4b65-b04d-50da92ee757b-kube-api-access-kxc2h\") pod \"barbican-7a2f-account-create-update-fjb4t\" (UID: \"fecdd245-3377-4b65-b04d-50da92ee757b\") " pod="openstack/barbican-7a2f-account-create-update-fjb4t" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.312253 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfbld\" (UniqueName: \"kubernetes.io/projected/2a4bd716-a01d-4e69-90ee-ae27b0e77c85-kube-api-access-qfbld\") pod \"barbican-db-create-rsc5q\" (UID: \"2a4bd716-a01d-4e69-90ee-ae27b0e77c85\") " pod="openstack/barbican-db-create-rsc5q" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.312384 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fecdd245-3377-4b65-b04d-50da92ee757b-operator-scripts\") pod \"barbican-7a2f-account-create-update-fjb4t\" (UID: \"fecdd245-3377-4b65-b04d-50da92ee757b\") " pod="openstack/barbican-7a2f-account-create-update-fjb4t" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.312520 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf7t9\" (UniqueName: \"kubernetes.io/projected/20531474-cb0d-4ce8-90fc-e9974acc6d1a-kube-api-access-wf7t9\") pod \"cinder-db-create-b4hsk\" (UID: \"20531474-cb0d-4ce8-90fc-e9974acc6d1a\") " pod="openstack/cinder-db-create-b4hsk" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.312637 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20531474-cb0d-4ce8-90fc-e9974acc6d1a-operator-scripts\") pod \"cinder-db-create-b4hsk\" (UID: \"20531474-cb0d-4ce8-90fc-e9974acc6d1a\") " pod="openstack/cinder-db-create-b4hsk" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.312781 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a4bd716-a01d-4e69-90ee-ae27b0e77c85-operator-scripts\") pod \"barbican-db-create-rsc5q\" (UID: \"2a4bd716-a01d-4e69-90ee-ae27b0e77c85\") " pod="openstack/barbican-db-create-rsc5q" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.313792 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a4bd716-a01d-4e69-90ee-ae27b0e77c85-operator-scripts\") pod \"barbican-db-create-rsc5q\" (UID: \"2a4bd716-a01d-4e69-90ee-ae27b0e77c85\") " pod="openstack/barbican-db-create-rsc5q" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.314530 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20531474-cb0d-4ce8-90fc-e9974acc6d1a-operator-scripts\") pod \"cinder-db-create-b4hsk\" (UID: \"20531474-cb0d-4ce8-90fc-e9974acc6d1a\") " pod="openstack/cinder-db-create-b4hsk" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.328539 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-7a2f-account-create-update-fjb4t"] Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.352482 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfbld\" (UniqueName: \"kubernetes.io/projected/2a4bd716-a01d-4e69-90ee-ae27b0e77c85-kube-api-access-qfbld\") pod \"barbican-db-create-rsc5q\" (UID: \"2a4bd716-a01d-4e69-90ee-ae27b0e77c85\") " pod="openstack/barbican-db-create-rsc5q" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.371264 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf7t9\" (UniqueName: \"kubernetes.io/projected/20531474-cb0d-4ce8-90fc-e9974acc6d1a-kube-api-access-wf7t9\") pod \"cinder-db-create-b4hsk\" (UID: \"20531474-cb0d-4ce8-90fc-e9974acc6d1a\") " pod="openstack/cinder-db-create-b4hsk" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.401355 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-7f14-account-create-update-kb9kq"] Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.402718 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7f14-account-create-update-kb9kq" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.407042 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.413887 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fecdd245-3377-4b65-b04d-50da92ee757b-operator-scripts\") pod \"barbican-7a2f-account-create-update-fjb4t\" (UID: \"fecdd245-3377-4b65-b04d-50da92ee757b\") " pod="openstack/barbican-7a2f-account-create-update-fjb4t" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.414238 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxc2h\" (UniqueName: \"kubernetes.io/projected/fecdd245-3377-4b65-b04d-50da92ee757b-kube-api-access-kxc2h\") pod \"barbican-7a2f-account-create-update-fjb4t\" (UID: \"fecdd245-3377-4b65-b04d-50da92ee757b\") " pod="openstack/barbican-7a2f-account-create-update-fjb4t" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.415447 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fecdd245-3377-4b65-b04d-50da92ee757b-operator-scripts\") pod \"barbican-7a2f-account-create-update-fjb4t\" (UID: \"fecdd245-3377-4b65-b04d-50da92ee757b\") " pod="openstack/barbican-7a2f-account-create-update-fjb4t" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.417493 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rsc5q" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.422645 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-7f14-account-create-update-kb9kq"] Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.480069 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxc2h\" (UniqueName: \"kubernetes.io/projected/fecdd245-3377-4b65-b04d-50da92ee757b-kube-api-access-kxc2h\") pod \"barbican-7a2f-account-create-update-fjb4t\" (UID: \"fecdd245-3377-4b65-b04d-50da92ee757b\") " pod="openstack/barbican-7a2f-account-create-update-fjb4t" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.493256 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-nj4w6"] Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.494393 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-nj4w6" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.502351 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.502861 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.502897 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ggfph" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.503315 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-b4hsk" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.513021 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-nj4w6"] Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.513130 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.515264 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4glv\" (UniqueName: \"kubernetes.io/projected/8cb17c41-9a33-45df-b013-044e94f413b4-kube-api-access-h4glv\") pod \"cinder-7f14-account-create-update-kb9kq\" (UID: \"8cb17c41-9a33-45df-b013-044e94f413b4\") " pod="openstack/cinder-7f14-account-create-update-kb9kq" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.515330 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8cb17c41-9a33-45df-b013-044e94f413b4-operator-scripts\") pod \"cinder-7f14-account-create-update-kb9kq\" (UID: \"8cb17c41-9a33-45df-b013-044e94f413b4\") " pod="openstack/cinder-7f14-account-create-update-kb9kq" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.613850 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7a2f-account-create-update-fjb4t" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.616462 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e11bf44e-799b-472b-aafb-89d5d071ad91-config-data\") pod \"keystone-db-sync-nj4w6\" (UID: \"e11bf44e-799b-472b-aafb-89d5d071ad91\") " pod="openstack/keystone-db-sync-nj4w6" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.616636 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4glv\" (UniqueName: \"kubernetes.io/projected/8cb17c41-9a33-45df-b013-044e94f413b4-kube-api-access-h4glv\") pod \"cinder-7f14-account-create-update-kb9kq\" (UID: \"8cb17c41-9a33-45df-b013-044e94f413b4\") " pod="openstack/cinder-7f14-account-create-update-kb9kq" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.616770 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e11bf44e-799b-472b-aafb-89d5d071ad91-combined-ca-bundle\") pod \"keystone-db-sync-nj4w6\" (UID: \"e11bf44e-799b-472b-aafb-89d5d071ad91\") " pod="openstack/keystone-db-sync-nj4w6" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.616854 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8cb17c41-9a33-45df-b013-044e94f413b4-operator-scripts\") pod \"cinder-7f14-account-create-update-kb9kq\" (UID: \"8cb17c41-9a33-45df-b013-044e94f413b4\") " pod="openstack/cinder-7f14-account-create-update-kb9kq" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.616989 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rx6kd\" (UniqueName: \"kubernetes.io/projected/e11bf44e-799b-472b-aafb-89d5d071ad91-kube-api-access-rx6kd\") pod \"keystone-db-sync-nj4w6\" (UID: \"e11bf44e-799b-472b-aafb-89d5d071ad91\") " pod="openstack/keystone-db-sync-nj4w6" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.618817 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8cb17c41-9a33-45df-b013-044e94f413b4-operator-scripts\") pod \"cinder-7f14-account-create-update-kb9kq\" (UID: \"8cb17c41-9a33-45df-b013-044e94f413b4\") " pod="openstack/cinder-7f14-account-create-update-kb9kq" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.635888 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4glv\" (UniqueName: \"kubernetes.io/projected/8cb17c41-9a33-45df-b013-044e94f413b4-kube-api-access-h4glv\") pod \"cinder-7f14-account-create-update-kb9kq\" (UID: \"8cb17c41-9a33-45df-b013-044e94f413b4\") " pod="openstack/cinder-7f14-account-create-update-kb9kq" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.718903 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e11bf44e-799b-472b-aafb-89d5d071ad91-combined-ca-bundle\") pod \"keystone-db-sync-nj4w6\" (UID: \"e11bf44e-799b-472b-aafb-89d5d071ad91\") " pod="openstack/keystone-db-sync-nj4w6" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.719024 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rx6kd\" (UniqueName: \"kubernetes.io/projected/e11bf44e-799b-472b-aafb-89d5d071ad91-kube-api-access-rx6kd\") pod \"keystone-db-sync-nj4w6\" (UID: \"e11bf44e-799b-472b-aafb-89d5d071ad91\") " pod="openstack/keystone-db-sync-nj4w6" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.719060 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e11bf44e-799b-472b-aafb-89d5d071ad91-config-data\") pod \"keystone-db-sync-nj4w6\" (UID: \"e11bf44e-799b-472b-aafb-89d5d071ad91\") " pod="openstack/keystone-db-sync-nj4w6" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.722166 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e11bf44e-799b-472b-aafb-89d5d071ad91-combined-ca-bundle\") pod \"keystone-db-sync-nj4w6\" (UID: \"e11bf44e-799b-472b-aafb-89d5d071ad91\") " pod="openstack/keystone-db-sync-nj4w6" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.725348 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e11bf44e-799b-472b-aafb-89d5d071ad91-config-data\") pod \"keystone-db-sync-nj4w6\" (UID: \"e11bf44e-799b-472b-aafb-89d5d071ad91\") " pod="openstack/keystone-db-sync-nj4w6" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.740136 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rx6kd\" (UniqueName: \"kubernetes.io/projected/e11bf44e-799b-472b-aafb-89d5d071ad91-kube-api-access-rx6kd\") pod \"keystone-db-sync-nj4w6\" (UID: \"e11bf44e-799b-472b-aafb-89d5d071ad91\") " pod="openstack/keystone-db-sync-nj4w6" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.847987 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7f14-account-create-update-kb9kq" Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.884477 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-nj4w6" Dec 03 16:23:22 crc kubenswrapper[4998]: W1203 16:23:22.985233 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a4bd716_a01d_4e69_90ee_ae27b0e77c85.slice/crio-ab6aa9b37d8ceb56f275ff3c31c763d559d8b80771f32085900129ba6ae6f44b WatchSource:0}: Error finding container ab6aa9b37d8ceb56f275ff3c31c763d559d8b80771f32085900129ba6ae6f44b: Status 404 returned error can't find the container with id ab6aa9b37d8ceb56f275ff3c31c763d559d8b80771f32085900129ba6ae6f44b Dec 03 16:23:22 crc kubenswrapper[4998]: I1203 16:23:22.992976 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-rsc5q"] Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.049296 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-b4hsk"] Dec 03 16:23:23 crc kubenswrapper[4998]: W1203 16:23:23.057102 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20531474_cb0d_4ce8_90fc_e9974acc6d1a.slice/crio-36d7060c99856eac8d1f501de32557e62a26bc0a1035d39a682707368886e062 WatchSource:0}: Error finding container 36d7060c99856eac8d1f501de32557e62a26bc0a1035d39a682707368886e062: Status 404 returned error can't find the container with id 36d7060c99856eac8d1f501de32557e62a26bc0a1035d39a682707368886e062 Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.122281 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-7a2f-account-create-update-fjb4t"] Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.167460 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-nj4w6"] Dec 03 16:23:23 crc kubenswrapper[4998]: W1203 16:23:23.177678 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode11bf44e_799b_472b_aafb_89d5d071ad91.slice/crio-486591e7c40536756dccef5049657a3f0842bd5227302a7c6351ebcf1195dbea WatchSource:0}: Error finding container 486591e7c40536756dccef5049657a3f0842bd5227302a7c6351ebcf1195dbea: Status 404 returned error can't find the container with id 486591e7c40536756dccef5049657a3f0842bd5227302a7c6351ebcf1195dbea Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.306470 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-7f14-account-create-update-kb9kq"] Dec 03 16:23:23 crc kubenswrapper[4998]: W1203 16:23:23.318692 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cb17c41_9a33_45df_b013_044e94f413b4.slice/crio-b8bbbe70ae2bd145accf5cac0815abe21aa5c5e8485059f9fb2f606a3356165b WatchSource:0}: Error finding container b8bbbe70ae2bd145accf5cac0815abe21aa5c5e8485059f9fb2f606a3356165b: Status 404 returned error can't find the container with id b8bbbe70ae2bd145accf5cac0815abe21aa5c5e8485059f9fb2f606a3356165b Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.350237 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-b4hsk" event={"ID":"20531474-cb0d-4ce8-90fc-e9974acc6d1a","Type":"ContainerStarted","Data":"2f36e3df4d17f0f527d02f6c2f1f136d16d80742bb0abcc3a1ebc80c76a50420"} Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.350289 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-b4hsk" event={"ID":"20531474-cb0d-4ce8-90fc-e9974acc6d1a","Type":"ContainerStarted","Data":"36d7060c99856eac8d1f501de32557e62a26bc0a1035d39a682707368886e062"} Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.353889 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rsc5q" event={"ID":"2a4bd716-a01d-4e69-90ee-ae27b0e77c85","Type":"ContainerStarted","Data":"02b6c984776989d01a404c0876ed79a235d0feb077b46a3356310c2a94dffa78"} Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.353939 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rsc5q" event={"ID":"2a4bd716-a01d-4e69-90ee-ae27b0e77c85","Type":"ContainerStarted","Data":"ab6aa9b37d8ceb56f275ff3c31c763d559d8b80771f32085900129ba6ae6f44b"} Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.360217 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7a2f-account-create-update-fjb4t" event={"ID":"fecdd245-3377-4b65-b04d-50da92ee757b","Type":"ContainerStarted","Data":"dcc8c7984741634c34bd1ede60a7504fc741bda719000c93a8ff3754c2e0c699"} Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.360261 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7a2f-account-create-update-fjb4t" event={"ID":"fecdd245-3377-4b65-b04d-50da92ee757b","Type":"ContainerStarted","Data":"6199177f80c59ad41f096e414286d8cf46cbd299bcde616f4ef08c2d1c810e99"} Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.367975 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-b4hsk" podStartSLOduration=1.367957894 podStartE2EDuration="1.367957894s" podCreationTimestamp="2025-12-03 16:23:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:23:23.363259166 +0000 UTC m=+1181.974959389" watchObservedRunningTime="2025-12-03 16:23:23.367957894 +0000 UTC m=+1181.979658117" Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.376180 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-nj4w6" event={"ID":"e11bf44e-799b-472b-aafb-89d5d071ad91","Type":"ContainerStarted","Data":"486591e7c40536756dccef5049657a3f0842bd5227302a7c6351ebcf1195dbea"} Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.377420 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-7f14-account-create-update-kb9kq" event={"ID":"8cb17c41-9a33-45df-b013-044e94f413b4","Type":"ContainerStarted","Data":"b8bbbe70ae2bd145accf5cac0815abe21aa5c5e8485059f9fb2f606a3356165b"} Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.406520 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-7a2f-account-create-update-fjb4t" podStartSLOduration=1.406506146 podStartE2EDuration="1.406506146s" podCreationTimestamp="2025-12-03 16:23:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:23:23.404722791 +0000 UTC m=+1182.016423014" watchObservedRunningTime="2025-12-03 16:23:23.406506146 +0000 UTC m=+1182.018206369" Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.408560 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-rsc5q" podStartSLOduration=1.408548757 podStartE2EDuration="1.408548757s" podCreationTimestamp="2025-12-03 16:23:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:23:23.390571544 +0000 UTC m=+1182.002271767" watchObservedRunningTime="2025-12-03 16:23:23.408548757 +0000 UTC m=+1182.020248980" Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.690085 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-cm9s2"] Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.691371 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-cm9s2" Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.696435 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-cm9s2"] Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.731259 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-db-sync-rjhmp"] Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.732242 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-rjhmp" Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.733848 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-config-data" Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.734069 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-watcher-dockercfg-vlg9j" Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.763669 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-sync-rjhmp"] Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.841119 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xmxh\" (UniqueName: \"kubernetes.io/projected/edc74770-171e-4ade-8dcd-bcaead186741-kube-api-access-4xmxh\") pod \"watcher-db-sync-rjhmp\" (UID: \"edc74770-171e-4ade-8dcd-bcaead186741\") " pod="openstack/watcher-db-sync-rjhmp" Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.842892 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/23518b26-e4ce-495d-aada-91f784244125-operator-scripts\") pod \"glance-db-create-cm9s2\" (UID: \"23518b26-e4ce-495d-aada-91f784244125\") " pod="openstack/glance-db-create-cm9s2" Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.843236 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/edc74770-171e-4ade-8dcd-bcaead186741-db-sync-config-data\") pod \"watcher-db-sync-rjhmp\" (UID: \"edc74770-171e-4ade-8dcd-bcaead186741\") " pod="openstack/watcher-db-sync-rjhmp" Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.843417 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edc74770-171e-4ade-8dcd-bcaead186741-config-data\") pod \"watcher-db-sync-rjhmp\" (UID: \"edc74770-171e-4ade-8dcd-bcaead186741\") " pod="openstack/watcher-db-sync-rjhmp" Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.843642 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edc74770-171e-4ade-8dcd-bcaead186741-combined-ca-bundle\") pod \"watcher-db-sync-rjhmp\" (UID: \"edc74770-171e-4ade-8dcd-bcaead186741\") " pod="openstack/watcher-db-sync-rjhmp" Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.844594 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqzs6\" (UniqueName: \"kubernetes.io/projected/23518b26-e4ce-495d-aada-91f784244125-kube-api-access-lqzs6\") pod \"glance-db-create-cm9s2\" (UID: \"23518b26-e4ce-495d-aada-91f784244125\") " pod="openstack/glance-db-create-cm9s2" Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.845552 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-ea32-account-create-update-zszrs"] Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.847860 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ea32-account-create-update-zszrs" Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.850626 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 03 16:23:23 crc kubenswrapper[4998]: I1203 16:23:23.872784 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-ea32-account-create-update-zszrs"] Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:23.945934 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/edc74770-171e-4ade-8dcd-bcaead186741-db-sync-config-data\") pod \"watcher-db-sync-rjhmp\" (UID: \"edc74770-171e-4ade-8dcd-bcaead186741\") " pod="openstack/watcher-db-sync-rjhmp" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:23.946010 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edc74770-171e-4ade-8dcd-bcaead186741-config-data\") pod \"watcher-db-sync-rjhmp\" (UID: \"edc74770-171e-4ade-8dcd-bcaead186741\") " pod="openstack/watcher-db-sync-rjhmp" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:23.946054 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edc74770-171e-4ade-8dcd-bcaead186741-combined-ca-bundle\") pod \"watcher-db-sync-rjhmp\" (UID: \"edc74770-171e-4ade-8dcd-bcaead186741\") " pod="openstack/watcher-db-sync-rjhmp" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:23.946085 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqzs6\" (UniqueName: \"kubernetes.io/projected/23518b26-e4ce-495d-aada-91f784244125-kube-api-access-lqzs6\") pod \"glance-db-create-cm9s2\" (UID: \"23518b26-e4ce-495d-aada-91f784244125\") " pod="openstack/glance-db-create-cm9s2" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:23.946131 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xmxh\" (UniqueName: \"kubernetes.io/projected/edc74770-171e-4ade-8dcd-bcaead186741-kube-api-access-4xmxh\") pod \"watcher-db-sync-rjhmp\" (UID: \"edc74770-171e-4ade-8dcd-bcaead186741\") " pod="openstack/watcher-db-sync-rjhmp" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:23.946164 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/23518b26-e4ce-495d-aada-91f784244125-operator-scripts\") pod \"glance-db-create-cm9s2\" (UID: \"23518b26-e4ce-495d-aada-91f784244125\") " pod="openstack/glance-db-create-cm9s2" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:23.947398 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/23518b26-e4ce-495d-aada-91f784244125-operator-scripts\") pod \"glance-db-create-cm9s2\" (UID: \"23518b26-e4ce-495d-aada-91f784244125\") " pod="openstack/glance-db-create-cm9s2" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:23.952084 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/edc74770-171e-4ade-8dcd-bcaead186741-db-sync-config-data\") pod \"watcher-db-sync-rjhmp\" (UID: \"edc74770-171e-4ade-8dcd-bcaead186741\") " pod="openstack/watcher-db-sync-rjhmp" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:23.952277 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edc74770-171e-4ade-8dcd-bcaead186741-config-data\") pod \"watcher-db-sync-rjhmp\" (UID: \"edc74770-171e-4ade-8dcd-bcaead186741\") " pod="openstack/watcher-db-sync-rjhmp" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:23.975664 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqzs6\" (UniqueName: \"kubernetes.io/projected/23518b26-e4ce-495d-aada-91f784244125-kube-api-access-lqzs6\") pod \"glance-db-create-cm9s2\" (UID: \"23518b26-e4ce-495d-aada-91f784244125\") " pod="openstack/glance-db-create-cm9s2" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:23.976098 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xmxh\" (UniqueName: \"kubernetes.io/projected/edc74770-171e-4ade-8dcd-bcaead186741-kube-api-access-4xmxh\") pod \"watcher-db-sync-rjhmp\" (UID: \"edc74770-171e-4ade-8dcd-bcaead186741\") " pod="openstack/watcher-db-sync-rjhmp" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:23.997208 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edc74770-171e-4ade-8dcd-bcaead186741-combined-ca-bundle\") pod \"watcher-db-sync-rjhmp\" (UID: \"edc74770-171e-4ade-8dcd-bcaead186741\") " pod="openstack/watcher-db-sync-rjhmp" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.003169 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-zv5gc"] Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.004593 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zv5gc" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.015245 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-zv5gc"] Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.029977 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-f9d0-account-create-update-pf87v"] Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.031076 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f9d0-account-create-update-pf87v" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.033174 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.048006 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f9d0-account-create-update-pf87v"] Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.052798 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8c2151f-9cdb-445e-865c-572da28ab46c-operator-scripts\") pod \"neutron-f9d0-account-create-update-pf87v\" (UID: \"c8c2151f-9cdb-445e-865c-572da28ab46c\") " pod="openstack/neutron-f9d0-account-create-update-pf87v" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.052844 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m96qs\" (UniqueName: \"kubernetes.io/projected/00a5807c-76d9-4a03-b8f1-1c8823915d32-kube-api-access-m96qs\") pod \"neutron-db-create-zv5gc\" (UID: \"00a5807c-76d9-4a03-b8f1-1c8823915d32\") " pod="openstack/neutron-db-create-zv5gc" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.052907 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8fc53f31-60a9-4a2d-9d61-6c3263f95ca9-operator-scripts\") pod \"glance-ea32-account-create-update-zszrs\" (UID: \"8fc53f31-60a9-4a2d-9d61-6c3263f95ca9\") " pod="openstack/glance-ea32-account-create-update-zszrs" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.052982 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcj9v\" (UniqueName: \"kubernetes.io/projected/c8c2151f-9cdb-445e-865c-572da28ab46c-kube-api-access-wcj9v\") pod \"neutron-f9d0-account-create-update-pf87v\" (UID: \"c8c2151f-9cdb-445e-865c-572da28ab46c\") " pod="openstack/neutron-f9d0-account-create-update-pf87v" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.053066 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00a5807c-76d9-4a03-b8f1-1c8823915d32-operator-scripts\") pod \"neutron-db-create-zv5gc\" (UID: \"00a5807c-76d9-4a03-b8f1-1c8823915d32\") " pod="openstack/neutron-db-create-zv5gc" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.053105 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psxpf\" (UniqueName: \"kubernetes.io/projected/8fc53f31-60a9-4a2d-9d61-6c3263f95ca9-kube-api-access-psxpf\") pod \"glance-ea32-account-create-update-zszrs\" (UID: \"8fc53f31-60a9-4a2d-9d61-6c3263f95ca9\") " pod="openstack/glance-ea32-account-create-update-zszrs" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.075335 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-cm9s2" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.084823 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-rjhmp" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.154888 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcj9v\" (UniqueName: \"kubernetes.io/projected/c8c2151f-9cdb-445e-865c-572da28ab46c-kube-api-access-wcj9v\") pod \"neutron-f9d0-account-create-update-pf87v\" (UID: \"c8c2151f-9cdb-445e-865c-572da28ab46c\") " pod="openstack/neutron-f9d0-account-create-update-pf87v" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.155176 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00a5807c-76d9-4a03-b8f1-1c8823915d32-operator-scripts\") pod \"neutron-db-create-zv5gc\" (UID: \"00a5807c-76d9-4a03-b8f1-1c8823915d32\") " pod="openstack/neutron-db-create-zv5gc" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.155207 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psxpf\" (UniqueName: \"kubernetes.io/projected/8fc53f31-60a9-4a2d-9d61-6c3263f95ca9-kube-api-access-psxpf\") pod \"glance-ea32-account-create-update-zszrs\" (UID: \"8fc53f31-60a9-4a2d-9d61-6c3263f95ca9\") " pod="openstack/glance-ea32-account-create-update-zszrs" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.155230 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8c2151f-9cdb-445e-865c-572da28ab46c-operator-scripts\") pod \"neutron-f9d0-account-create-update-pf87v\" (UID: \"c8c2151f-9cdb-445e-865c-572da28ab46c\") " pod="openstack/neutron-f9d0-account-create-update-pf87v" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.155265 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m96qs\" (UniqueName: \"kubernetes.io/projected/00a5807c-76d9-4a03-b8f1-1c8823915d32-kube-api-access-m96qs\") pod \"neutron-db-create-zv5gc\" (UID: \"00a5807c-76d9-4a03-b8f1-1c8823915d32\") " pod="openstack/neutron-db-create-zv5gc" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.155319 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8fc53f31-60a9-4a2d-9d61-6c3263f95ca9-operator-scripts\") pod \"glance-ea32-account-create-update-zszrs\" (UID: \"8fc53f31-60a9-4a2d-9d61-6c3263f95ca9\") " pod="openstack/glance-ea32-account-create-update-zszrs" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.156073 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00a5807c-76d9-4a03-b8f1-1c8823915d32-operator-scripts\") pod \"neutron-db-create-zv5gc\" (UID: \"00a5807c-76d9-4a03-b8f1-1c8823915d32\") " pod="openstack/neutron-db-create-zv5gc" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.156080 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8fc53f31-60a9-4a2d-9d61-6c3263f95ca9-operator-scripts\") pod \"glance-ea32-account-create-update-zszrs\" (UID: \"8fc53f31-60a9-4a2d-9d61-6c3263f95ca9\") " pod="openstack/glance-ea32-account-create-update-zszrs" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.156525 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8c2151f-9cdb-445e-865c-572da28ab46c-operator-scripts\") pod \"neutron-f9d0-account-create-update-pf87v\" (UID: \"c8c2151f-9cdb-445e-865c-572da28ab46c\") " pod="openstack/neutron-f9d0-account-create-update-pf87v" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.179357 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcj9v\" (UniqueName: \"kubernetes.io/projected/c8c2151f-9cdb-445e-865c-572da28ab46c-kube-api-access-wcj9v\") pod \"neutron-f9d0-account-create-update-pf87v\" (UID: \"c8c2151f-9cdb-445e-865c-572da28ab46c\") " pod="openstack/neutron-f9d0-account-create-update-pf87v" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.179362 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psxpf\" (UniqueName: \"kubernetes.io/projected/8fc53f31-60a9-4a2d-9d61-6c3263f95ca9-kube-api-access-psxpf\") pod \"glance-ea32-account-create-update-zszrs\" (UID: \"8fc53f31-60a9-4a2d-9d61-6c3263f95ca9\") " pod="openstack/glance-ea32-account-create-update-zszrs" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.180935 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m96qs\" (UniqueName: \"kubernetes.io/projected/00a5807c-76d9-4a03-b8f1-1c8823915d32-kube-api-access-m96qs\") pod \"neutron-db-create-zv5gc\" (UID: \"00a5807c-76d9-4a03-b8f1-1c8823915d32\") " pod="openstack/neutron-db-create-zv5gc" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.373332 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zv5gc" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.383839 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f9d0-account-create-update-pf87v" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.423639 4998 generic.go:334] "Generic (PLEG): container finished" podID="8cb17c41-9a33-45df-b013-044e94f413b4" containerID="854db052dba6cd5655d3c5d1c22c7dd1fd27a486305b18052c7f7d315ddf755d" exitCode=0 Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.424044 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-7f14-account-create-update-kb9kq" event={"ID":"8cb17c41-9a33-45df-b013-044e94f413b4","Type":"ContainerDied","Data":"854db052dba6cd5655d3c5d1c22c7dd1fd27a486305b18052c7f7d315ddf755d"} Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.435409 4998 generic.go:334] "Generic (PLEG): container finished" podID="20531474-cb0d-4ce8-90fc-e9974acc6d1a" containerID="2f36e3df4d17f0f527d02f6c2f1f136d16d80742bb0abcc3a1ebc80c76a50420" exitCode=0 Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.435598 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-b4hsk" event={"ID":"20531474-cb0d-4ce8-90fc-e9974acc6d1a","Type":"ContainerDied","Data":"2f36e3df4d17f0f527d02f6c2f1f136d16d80742bb0abcc3a1ebc80c76a50420"} Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.452014 4998 generic.go:334] "Generic (PLEG): container finished" podID="2a4bd716-a01d-4e69-90ee-ae27b0e77c85" containerID="02b6c984776989d01a404c0876ed79a235d0feb077b46a3356310c2a94dffa78" exitCode=0 Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.452091 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rsc5q" event={"ID":"2a4bd716-a01d-4e69-90ee-ae27b0e77c85","Type":"ContainerDied","Data":"02b6c984776989d01a404c0876ed79a235d0feb077b46a3356310c2a94dffa78"} Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.482554 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ea32-account-create-update-zszrs" Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.483814 4998 generic.go:334] "Generic (PLEG): container finished" podID="fecdd245-3377-4b65-b04d-50da92ee757b" containerID="dcc8c7984741634c34bd1ede60a7504fc741bda719000c93a8ff3754c2e0c699" exitCode=0 Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.483889 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7a2f-account-create-update-fjb4t" event={"ID":"fecdd245-3377-4b65-b04d-50da92ee757b","Type":"ContainerDied","Data":"dcc8c7984741634c34bd1ede60a7504fc741bda719000c93a8ff3754c2e0c699"} Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.578140 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-cm9s2"] Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.668030 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-sync-rjhmp"] Dec 03 16:23:24 crc kubenswrapper[4998]: W1203 16:23:24.752151 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podedc74770_171e_4ade_8dcd_bcaead186741.slice/crio-e001f3a8a1b54ab36a1848a785335e49a5222831bbedf30382e2e8ae6f6ec6b7 WatchSource:0}: Error finding container e001f3a8a1b54ab36a1848a785335e49a5222831bbedf30382e2e8ae6f6ec6b7: Status 404 returned error can't find the container with id e001f3a8a1b54ab36a1848a785335e49a5222831bbedf30382e2e8ae6f6ec6b7 Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.931090 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-ea32-account-create-update-zszrs"] Dec 03 16:23:24 crc kubenswrapper[4998]: W1203 16:23:24.936659 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8fc53f31_60a9_4a2d_9d61_6c3263f95ca9.slice/crio-df46be464bd4f4cbb403169207fbdee564f2f0be5ff150e64bec488d0ef4a4f6 WatchSource:0}: Error finding container df46be464bd4f4cbb403169207fbdee564f2f0be5ff150e64bec488d0ef4a4f6: Status 404 returned error can't find the container with id df46be464bd4f4cbb403169207fbdee564f2f0be5ff150e64bec488d0ef4a4f6 Dec 03 16:23:24 crc kubenswrapper[4998]: I1203 16:23:24.979710 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f9d0-account-create-update-pf87v"] Dec 03 16:23:25 crc kubenswrapper[4998]: I1203 16:23:25.013653 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-zv5gc"] Dec 03 16:23:25 crc kubenswrapper[4998]: W1203 16:23:25.027478 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00a5807c_76d9_4a03_b8f1_1c8823915d32.slice/crio-6522e0b3bb71f95ba2e8f493808bc68c8b1f67b30af24efd7cee1ebc5399b4df WatchSource:0}: Error finding container 6522e0b3bb71f95ba2e8f493808bc68c8b1f67b30af24efd7cee1ebc5399b4df: Status 404 returned error can't find the container with id 6522e0b3bb71f95ba2e8f493808bc68c8b1f67b30af24efd7cee1ebc5399b4df Dec 03 16:23:25 crc kubenswrapper[4998]: I1203 16:23:25.495490 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-zv5gc" event={"ID":"00a5807c-76d9-4a03-b8f1-1c8823915d32","Type":"ContainerStarted","Data":"50b57c87c0146a67455069dc0f21e106f57997c31cb455ec6bab3e1ee1d4a159"} Dec 03 16:23:25 crc kubenswrapper[4998]: I1203 16:23:25.495696 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-zv5gc" event={"ID":"00a5807c-76d9-4a03-b8f1-1c8823915d32","Type":"ContainerStarted","Data":"6522e0b3bb71f95ba2e8f493808bc68c8b1f67b30af24efd7cee1ebc5399b4df"} Dec 03 16:23:25 crc kubenswrapper[4998]: I1203 16:23:25.500511 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f9d0-account-create-update-pf87v" event={"ID":"c8c2151f-9cdb-445e-865c-572da28ab46c","Type":"ContainerStarted","Data":"66b40d6d558616b9828bc83b8a9c16287350b7bba9b9fcc34de50c5dfa9e4466"} Dec 03 16:23:25 crc kubenswrapper[4998]: I1203 16:23:25.500559 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f9d0-account-create-update-pf87v" event={"ID":"c8c2151f-9cdb-445e-865c-572da28ab46c","Type":"ContainerStarted","Data":"cabc8e29e8e133c328e662509e9fcbd1b9e4dedf3df3b72a30275d5f3f3201b7"} Dec 03 16:23:25 crc kubenswrapper[4998]: I1203 16:23:25.502458 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ea32-account-create-update-zszrs" event={"ID":"8fc53f31-60a9-4a2d-9d61-6c3263f95ca9","Type":"ContainerStarted","Data":"e8281d3840790c2c034b305a7252b0b14cc1498567f5d7cb52a02564cb924e6d"} Dec 03 16:23:25 crc kubenswrapper[4998]: I1203 16:23:25.502489 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ea32-account-create-update-zszrs" event={"ID":"8fc53f31-60a9-4a2d-9d61-6c3263f95ca9","Type":"ContainerStarted","Data":"df46be464bd4f4cbb403169207fbdee564f2f0be5ff150e64bec488d0ef4a4f6"} Dec 03 16:23:25 crc kubenswrapper[4998]: I1203 16:23:25.503859 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-rjhmp" event={"ID":"edc74770-171e-4ade-8dcd-bcaead186741","Type":"ContainerStarted","Data":"e001f3a8a1b54ab36a1848a785335e49a5222831bbedf30382e2e8ae6f6ec6b7"} Dec 03 16:23:25 crc kubenswrapper[4998]: I1203 16:23:25.505309 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-cm9s2" event={"ID":"23518b26-e4ce-495d-aada-91f784244125","Type":"ContainerStarted","Data":"d95b944c60c5729c15567f7c9cd0d472ca7f9356c0a1849a2e87b6e9cef89446"} Dec 03 16:23:25 crc kubenswrapper[4998]: I1203 16:23:25.505351 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-cm9s2" event={"ID":"23518b26-e4ce-495d-aada-91f784244125","Type":"ContainerStarted","Data":"6b50d29397d65f35244b0962330672cbdd9a80eab224c60e21f82e4c2fbb7ae9"} Dec 03 16:23:25 crc kubenswrapper[4998]: I1203 16:23:25.537267 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-cm9s2" podStartSLOduration=2.537245616 podStartE2EDuration="2.537245616s" podCreationTimestamp="2025-12-03 16:23:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:23:25.53621529 +0000 UTC m=+1184.147915513" watchObservedRunningTime="2025-12-03 16:23:25.537245616 +0000 UTC m=+1184.148945839" Dec 03 16:23:25 crc kubenswrapper[4998]: I1203 16:23:25.544791 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-zv5gc" podStartSLOduration=2.544769436 podStartE2EDuration="2.544769436s" podCreationTimestamp="2025-12-03 16:23:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:23:25.522131495 +0000 UTC m=+1184.133831728" watchObservedRunningTime="2025-12-03 16:23:25.544769436 +0000 UTC m=+1184.156469659" Dec 03 16:23:25 crc kubenswrapper[4998]: I1203 16:23:25.557479 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-ea32-account-create-update-zszrs" podStartSLOduration=2.557459915 podStartE2EDuration="2.557459915s" podCreationTimestamp="2025-12-03 16:23:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:23:25.553941837 +0000 UTC m=+1184.165642050" watchObservedRunningTime="2025-12-03 16:23:25.557459915 +0000 UTC m=+1184.169160138" Dec 03 16:23:25 crc kubenswrapper[4998]: I1203 16:23:25.577472 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-f9d0-account-create-update-pf87v" podStartSLOduration=2.577454529 podStartE2EDuration="2.577454529s" podCreationTimestamp="2025-12-03 16:23:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:23:25.568856413 +0000 UTC m=+1184.180556636" watchObservedRunningTime="2025-12-03 16:23:25.577454529 +0000 UTC m=+1184.189154752" Dec 03 16:23:26 crc kubenswrapper[4998]: I1203 16:23:26.518818 4998 generic.go:334] "Generic (PLEG): container finished" podID="23518b26-e4ce-495d-aada-91f784244125" containerID="d95b944c60c5729c15567f7c9cd0d472ca7f9356c0a1849a2e87b6e9cef89446" exitCode=0 Dec 03 16:23:26 crc kubenswrapper[4998]: I1203 16:23:26.518881 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-cm9s2" event={"ID":"23518b26-e4ce-495d-aada-91f784244125","Type":"ContainerDied","Data":"d95b944c60c5729c15567f7c9cd0d472ca7f9356c0a1849a2e87b6e9cef89446"} Dec 03 16:23:26 crc kubenswrapper[4998]: I1203 16:23:26.521598 4998 generic.go:334] "Generic (PLEG): container finished" podID="00a5807c-76d9-4a03-b8f1-1c8823915d32" containerID="50b57c87c0146a67455069dc0f21e106f57997c31cb455ec6bab3e1ee1d4a159" exitCode=0 Dec 03 16:23:26 crc kubenswrapper[4998]: I1203 16:23:26.521679 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-zv5gc" event={"ID":"00a5807c-76d9-4a03-b8f1-1c8823915d32","Type":"ContainerDied","Data":"50b57c87c0146a67455069dc0f21e106f57997c31cb455ec6bab3e1ee1d4a159"} Dec 03 16:23:26 crc kubenswrapper[4998]: I1203 16:23:26.523847 4998 generic.go:334] "Generic (PLEG): container finished" podID="c8c2151f-9cdb-445e-865c-572da28ab46c" containerID="66b40d6d558616b9828bc83b8a9c16287350b7bba9b9fcc34de50c5dfa9e4466" exitCode=0 Dec 03 16:23:26 crc kubenswrapper[4998]: I1203 16:23:26.523920 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f9d0-account-create-update-pf87v" event={"ID":"c8c2151f-9cdb-445e-865c-572da28ab46c","Type":"ContainerDied","Data":"66b40d6d558616b9828bc83b8a9c16287350b7bba9b9fcc34de50c5dfa9e4466"} Dec 03 16:23:26 crc kubenswrapper[4998]: I1203 16:23:26.526282 4998 generic.go:334] "Generic (PLEG): container finished" podID="8fc53f31-60a9-4a2d-9d61-6c3263f95ca9" containerID="e8281d3840790c2c034b305a7252b0b14cc1498567f5d7cb52a02564cb924e6d" exitCode=0 Dec 03 16:23:26 crc kubenswrapper[4998]: I1203 16:23:26.526335 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ea32-account-create-update-zszrs" event={"ID":"8fc53f31-60a9-4a2d-9d61-6c3263f95ca9","Type":"ContainerDied","Data":"e8281d3840790c2c034b305a7252b0b14cc1498567f5d7cb52a02564cb924e6d"} Dec 03 16:23:26 crc kubenswrapper[4998]: I1203 16:23:26.983953 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-946dbfbcf-7n244" Dec 03 16:23:27 crc kubenswrapper[4998]: I1203 16:23:27.040454 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55ddfd5dfc-9clrt"] Dec 03 16:23:27 crc kubenswrapper[4998]: I1203 16:23:27.040720 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" podUID="e266f963-33c9-48dc-bb10-70bd5c700d69" containerName="dnsmasq-dns" containerID="cri-o://8a02ff22240d153f314a59c60dd4559c01cd622cfce7f13356e6514f03fde803" gracePeriod=10 Dec 03 16:23:27 crc kubenswrapper[4998]: I1203 16:23:27.136167 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" podUID="e266f963-33c9-48dc-bb10-70bd5c700d69" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.121:5353: connect: connection refused" Dec 03 16:23:27 crc kubenswrapper[4998]: I1203 16:23:27.542092 4998 generic.go:334] "Generic (PLEG): container finished" podID="a70537b1-93bc-48c3-9bef-fd9030158cad" containerID="9bd7f454c517ad30a0d2fb22370f32404af1a46c10a6bbd97476e74751b2c2c1" exitCode=0 Dec 03 16:23:27 crc kubenswrapper[4998]: I1203 16:23:27.542341 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a70537b1-93bc-48c3-9bef-fd9030158cad","Type":"ContainerDied","Data":"9bd7f454c517ad30a0d2fb22370f32404af1a46c10a6bbd97476e74751b2c2c1"} Dec 03 16:23:29 crc kubenswrapper[4998]: I1203 16:23:29.564517 4998 generic.go:334] "Generic (PLEG): container finished" podID="e266f963-33c9-48dc-bb10-70bd5c700d69" containerID="8a02ff22240d153f314a59c60dd4559c01cd622cfce7f13356e6514f03fde803" exitCode=0 Dec 03 16:23:29 crc kubenswrapper[4998]: I1203 16:23:29.564615 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" event={"ID":"e266f963-33c9-48dc-bb10-70bd5c700d69","Type":"ContainerDied","Data":"8a02ff22240d153f314a59c60dd4559c01cd622cfce7f13356e6514f03fde803"} Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.249446 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-b4hsk" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.256509 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rsc5q" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.279873 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7a2f-account-create-update-fjb4t" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.294155 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7f14-account-create-update-kb9kq" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.299585 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zv5gc" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.300032 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ea32-account-create-update-zszrs" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.313255 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-cm9s2" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.326939 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f9d0-account-create-update-pf87v" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.395045 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfbld\" (UniqueName: \"kubernetes.io/projected/2a4bd716-a01d-4e69-90ee-ae27b0e77c85-kube-api-access-qfbld\") pod \"2a4bd716-a01d-4e69-90ee-ae27b0e77c85\" (UID: \"2a4bd716-a01d-4e69-90ee-ae27b0e77c85\") " Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.395109 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxc2h\" (UniqueName: \"kubernetes.io/projected/fecdd245-3377-4b65-b04d-50da92ee757b-kube-api-access-kxc2h\") pod \"fecdd245-3377-4b65-b04d-50da92ee757b\" (UID: \"fecdd245-3377-4b65-b04d-50da92ee757b\") " Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.395127 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8fc53f31-60a9-4a2d-9d61-6c3263f95ca9-operator-scripts\") pod \"8fc53f31-60a9-4a2d-9d61-6c3263f95ca9\" (UID: \"8fc53f31-60a9-4a2d-9d61-6c3263f95ca9\") " Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.395188 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00a5807c-76d9-4a03-b8f1-1c8823915d32-operator-scripts\") pod \"00a5807c-76d9-4a03-b8f1-1c8823915d32\" (UID: \"00a5807c-76d9-4a03-b8f1-1c8823915d32\") " Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.395220 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/23518b26-e4ce-495d-aada-91f784244125-operator-scripts\") pod \"23518b26-e4ce-495d-aada-91f784244125\" (UID: \"23518b26-e4ce-495d-aada-91f784244125\") " Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.395238 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20531474-cb0d-4ce8-90fc-e9974acc6d1a-operator-scripts\") pod \"20531474-cb0d-4ce8-90fc-e9974acc6d1a\" (UID: \"20531474-cb0d-4ce8-90fc-e9974acc6d1a\") " Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.395317 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psxpf\" (UniqueName: \"kubernetes.io/projected/8fc53f31-60a9-4a2d-9d61-6c3263f95ca9-kube-api-access-psxpf\") pod \"8fc53f31-60a9-4a2d-9d61-6c3263f95ca9\" (UID: \"8fc53f31-60a9-4a2d-9d61-6c3263f95ca9\") " Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.395336 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m96qs\" (UniqueName: \"kubernetes.io/projected/00a5807c-76d9-4a03-b8f1-1c8823915d32-kube-api-access-m96qs\") pod \"00a5807c-76d9-4a03-b8f1-1c8823915d32\" (UID: \"00a5807c-76d9-4a03-b8f1-1c8823915d32\") " Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.395368 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fecdd245-3377-4b65-b04d-50da92ee757b-operator-scripts\") pod \"fecdd245-3377-4b65-b04d-50da92ee757b\" (UID: \"fecdd245-3377-4b65-b04d-50da92ee757b\") " Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.395392 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqzs6\" (UniqueName: \"kubernetes.io/projected/23518b26-e4ce-495d-aada-91f784244125-kube-api-access-lqzs6\") pod \"23518b26-e4ce-495d-aada-91f784244125\" (UID: \"23518b26-e4ce-495d-aada-91f784244125\") " Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.395420 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wf7t9\" (UniqueName: \"kubernetes.io/projected/20531474-cb0d-4ce8-90fc-e9974acc6d1a-kube-api-access-wf7t9\") pod \"20531474-cb0d-4ce8-90fc-e9974acc6d1a\" (UID: \"20531474-cb0d-4ce8-90fc-e9974acc6d1a\") " Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.395444 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a4bd716-a01d-4e69-90ee-ae27b0e77c85-operator-scripts\") pod \"2a4bd716-a01d-4e69-90ee-ae27b0e77c85\" (UID: \"2a4bd716-a01d-4e69-90ee-ae27b0e77c85\") " Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.395480 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8cb17c41-9a33-45df-b013-044e94f413b4-operator-scripts\") pod \"8cb17c41-9a33-45df-b013-044e94f413b4\" (UID: \"8cb17c41-9a33-45df-b013-044e94f413b4\") " Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.395501 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4glv\" (UniqueName: \"kubernetes.io/projected/8cb17c41-9a33-45df-b013-044e94f413b4-kube-api-access-h4glv\") pod \"8cb17c41-9a33-45df-b013-044e94f413b4\" (UID: \"8cb17c41-9a33-45df-b013-044e94f413b4\") " Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.396479 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23518b26-e4ce-495d-aada-91f784244125-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "23518b26-e4ce-495d-aada-91f784244125" (UID: "23518b26-e4ce-495d-aada-91f784244125"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.396678 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cb17c41-9a33-45df-b013-044e94f413b4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8cb17c41-9a33-45df-b013-044e94f413b4" (UID: "8cb17c41-9a33-45df-b013-044e94f413b4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.396713 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00a5807c-76d9-4a03-b8f1-1c8823915d32-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "00a5807c-76d9-4a03-b8f1-1c8823915d32" (UID: "00a5807c-76d9-4a03-b8f1-1c8823915d32"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.396512 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20531474-cb0d-4ce8-90fc-e9974acc6d1a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "20531474-cb0d-4ce8-90fc-e9974acc6d1a" (UID: "20531474-cb0d-4ce8-90fc-e9974acc6d1a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.396986 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a4bd716-a01d-4e69-90ee-ae27b0e77c85-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2a4bd716-a01d-4e69-90ee-ae27b0e77c85" (UID: "2a4bd716-a01d-4e69-90ee-ae27b0e77c85"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.396967 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fecdd245-3377-4b65-b04d-50da92ee757b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fecdd245-3377-4b65-b04d-50da92ee757b" (UID: "fecdd245-3377-4b65-b04d-50da92ee757b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.397678 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fc53f31-60a9-4a2d-9d61-6c3263f95ca9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8fc53f31-60a9-4a2d-9d61-6c3263f95ca9" (UID: "8fc53f31-60a9-4a2d-9d61-6c3263f95ca9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.402199 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fecdd245-3377-4b65-b04d-50da92ee757b-kube-api-access-kxc2h" (OuterVolumeSpecName: "kube-api-access-kxc2h") pod "fecdd245-3377-4b65-b04d-50da92ee757b" (UID: "fecdd245-3377-4b65-b04d-50da92ee757b"). InnerVolumeSpecName "kube-api-access-kxc2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.402261 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cb17c41-9a33-45df-b013-044e94f413b4-kube-api-access-h4glv" (OuterVolumeSpecName: "kube-api-access-h4glv") pod "8cb17c41-9a33-45df-b013-044e94f413b4" (UID: "8cb17c41-9a33-45df-b013-044e94f413b4"). InnerVolumeSpecName "kube-api-access-h4glv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.402497 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00a5807c-76d9-4a03-b8f1-1c8823915d32-kube-api-access-m96qs" (OuterVolumeSpecName: "kube-api-access-m96qs") pod "00a5807c-76d9-4a03-b8f1-1c8823915d32" (UID: "00a5807c-76d9-4a03-b8f1-1c8823915d32"). InnerVolumeSpecName "kube-api-access-m96qs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.402631 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23518b26-e4ce-495d-aada-91f784244125-kube-api-access-lqzs6" (OuterVolumeSpecName: "kube-api-access-lqzs6") pod "23518b26-e4ce-495d-aada-91f784244125" (UID: "23518b26-e4ce-495d-aada-91f784244125"). InnerVolumeSpecName "kube-api-access-lqzs6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.402802 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20531474-cb0d-4ce8-90fc-e9974acc6d1a-kube-api-access-wf7t9" (OuterVolumeSpecName: "kube-api-access-wf7t9") pod "20531474-cb0d-4ce8-90fc-e9974acc6d1a" (UID: "20531474-cb0d-4ce8-90fc-e9974acc6d1a"). InnerVolumeSpecName "kube-api-access-wf7t9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.404068 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fc53f31-60a9-4a2d-9d61-6c3263f95ca9-kube-api-access-psxpf" (OuterVolumeSpecName: "kube-api-access-psxpf") pod "8fc53f31-60a9-4a2d-9d61-6c3263f95ca9" (UID: "8fc53f31-60a9-4a2d-9d61-6c3263f95ca9"). InnerVolumeSpecName "kube-api-access-psxpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.413882 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a4bd716-a01d-4e69-90ee-ae27b0e77c85-kube-api-access-qfbld" (OuterVolumeSpecName: "kube-api-access-qfbld") pod "2a4bd716-a01d-4e69-90ee-ae27b0e77c85" (UID: "2a4bd716-a01d-4e69-90ee-ae27b0e77c85"). InnerVolumeSpecName "kube-api-access-qfbld". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.497333 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8c2151f-9cdb-445e-865c-572da28ab46c-operator-scripts\") pod \"c8c2151f-9cdb-445e-865c-572da28ab46c\" (UID: \"c8c2151f-9cdb-445e-865c-572da28ab46c\") " Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.497402 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcj9v\" (UniqueName: \"kubernetes.io/projected/c8c2151f-9cdb-445e-865c-572da28ab46c-kube-api-access-wcj9v\") pod \"c8c2151f-9cdb-445e-865c-572da28ab46c\" (UID: \"c8c2151f-9cdb-445e-865c-572da28ab46c\") " Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.497763 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8c2151f-9cdb-445e-865c-572da28ab46c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c8c2151f-9cdb-445e-865c-572da28ab46c" (UID: "c8c2151f-9cdb-445e-865c-572da28ab46c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.497804 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfbld\" (UniqueName: \"kubernetes.io/projected/2a4bd716-a01d-4e69-90ee-ae27b0e77c85-kube-api-access-qfbld\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.497816 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxc2h\" (UniqueName: \"kubernetes.io/projected/fecdd245-3377-4b65-b04d-50da92ee757b-kube-api-access-kxc2h\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.497826 4998 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8fc53f31-60a9-4a2d-9d61-6c3263f95ca9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.497836 4998 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00a5807c-76d9-4a03-b8f1-1c8823915d32-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.497844 4998 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20531474-cb0d-4ce8-90fc-e9974acc6d1a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.497852 4998 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/23518b26-e4ce-495d-aada-91f784244125-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.497861 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-psxpf\" (UniqueName: \"kubernetes.io/projected/8fc53f31-60a9-4a2d-9d61-6c3263f95ca9-kube-api-access-psxpf\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.497869 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m96qs\" (UniqueName: \"kubernetes.io/projected/00a5807c-76d9-4a03-b8f1-1c8823915d32-kube-api-access-m96qs\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.497877 4998 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fecdd245-3377-4b65-b04d-50da92ee757b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.497886 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqzs6\" (UniqueName: \"kubernetes.io/projected/23518b26-e4ce-495d-aada-91f784244125-kube-api-access-lqzs6\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.497895 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wf7t9\" (UniqueName: \"kubernetes.io/projected/20531474-cb0d-4ce8-90fc-e9974acc6d1a-kube-api-access-wf7t9\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.497902 4998 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a4bd716-a01d-4e69-90ee-ae27b0e77c85-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.497910 4998 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8cb17c41-9a33-45df-b013-044e94f413b4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.497919 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4glv\" (UniqueName: \"kubernetes.io/projected/8cb17c41-9a33-45df-b013-044e94f413b4-kube-api-access-h4glv\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.500486 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8c2151f-9cdb-445e-865c-572da28ab46c-kube-api-access-wcj9v" (OuterVolumeSpecName: "kube-api-access-wcj9v") pod "c8c2151f-9cdb-445e-865c-572da28ab46c" (UID: "c8c2151f-9cdb-445e-865c-572da28ab46c"). InnerVolumeSpecName "kube-api-access-wcj9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.573106 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zv5gc" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.573770 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-zv5gc" event={"ID":"00a5807c-76d9-4a03-b8f1-1c8823915d32","Type":"ContainerDied","Data":"6522e0b3bb71f95ba2e8f493808bc68c8b1f67b30af24efd7cee1ebc5399b4df"} Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.573802 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6522e0b3bb71f95ba2e8f493808bc68c8b1f67b30af24efd7cee1ebc5399b4df" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.575210 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-7f14-account-create-update-kb9kq" event={"ID":"8cb17c41-9a33-45df-b013-044e94f413b4","Type":"ContainerDied","Data":"b8bbbe70ae2bd145accf5cac0815abe21aa5c5e8485059f9fb2f606a3356165b"} Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.575229 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8bbbe70ae2bd145accf5cac0815abe21aa5c5e8485059f9fb2f606a3356165b" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.575271 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-7f14-account-create-update-kb9kq" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.582561 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f9d0-account-create-update-pf87v" event={"ID":"c8c2151f-9cdb-445e-865c-572da28ab46c","Type":"ContainerDied","Data":"cabc8e29e8e133c328e662509e9fcbd1b9e4dedf3df3b72a30275d5f3f3201b7"} Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.582608 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f9d0-account-create-update-pf87v" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.582621 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cabc8e29e8e133c328e662509e9fcbd1b9e4dedf3df3b72a30275d5f3f3201b7" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.584302 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rsc5q" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.584396 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rsc5q" event={"ID":"2a4bd716-a01d-4e69-90ee-ae27b0e77c85","Type":"ContainerDied","Data":"ab6aa9b37d8ceb56f275ff3c31c763d559d8b80771f32085900129ba6ae6f44b"} Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.584510 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab6aa9b37d8ceb56f275ff3c31c763d559d8b80771f32085900129ba6ae6f44b" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.585883 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7a2f-account-create-update-fjb4t" event={"ID":"fecdd245-3377-4b65-b04d-50da92ee757b","Type":"ContainerDied","Data":"6199177f80c59ad41f096e414286d8cf46cbd299bcde616f4ef08c2d1c810e99"} Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.585906 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6199177f80c59ad41f096e414286d8cf46cbd299bcde616f4ef08c2d1c810e99" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.585946 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7a2f-account-create-update-fjb4t" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.588035 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ea32-account-create-update-zszrs" event={"ID":"8fc53f31-60a9-4a2d-9d61-6c3263f95ca9","Type":"ContainerDied","Data":"df46be464bd4f4cbb403169207fbdee564f2f0be5ff150e64bec488d0ef4a4f6"} Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.588087 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df46be464bd4f4cbb403169207fbdee564f2f0be5ff150e64bec488d0ef4a4f6" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.588151 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ea32-account-create-update-zszrs" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.590911 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-cm9s2" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.590909 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-cm9s2" event={"ID":"23518b26-e4ce-495d-aada-91f784244125","Type":"ContainerDied","Data":"6b50d29397d65f35244b0962330672cbdd9a80eab224c60e21f82e4c2fbb7ae9"} Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.591159 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b50d29397d65f35244b0962330672cbdd9a80eab224c60e21f82e4c2fbb7ae9" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.592936 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-b4hsk" event={"ID":"20531474-cb0d-4ce8-90fc-e9974acc6d1a","Type":"ContainerDied","Data":"36d7060c99856eac8d1f501de32557e62a26bc0a1035d39a682707368886e062"} Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.592959 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36d7060c99856eac8d1f501de32557e62a26bc0a1035d39a682707368886e062" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.592994 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-b4hsk" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.599514 4998 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c8c2151f-9cdb-445e-865c-572da28ab46c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:30 crc kubenswrapper[4998]: I1203 16:23:30.599558 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcj9v\" (UniqueName: \"kubernetes.io/projected/c8c2151f-9cdb-445e-865c-572da28ab46c-kube-api-access-wcj9v\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.029698 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-2rqhm"] Dec 03 16:23:34 crc kubenswrapper[4998]: E1203 16:23:34.031044 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a4bd716-a01d-4e69-90ee-ae27b0e77c85" containerName="mariadb-database-create" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.031136 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a4bd716-a01d-4e69-90ee-ae27b0e77c85" containerName="mariadb-database-create" Dec 03 16:23:34 crc kubenswrapper[4998]: E1203 16:23:34.031205 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cb17c41-9a33-45df-b013-044e94f413b4" containerName="mariadb-account-create-update" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.031269 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cb17c41-9a33-45df-b013-044e94f413b4" containerName="mariadb-account-create-update" Dec 03 16:23:34 crc kubenswrapper[4998]: E1203 16:23:34.031346 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8c2151f-9cdb-445e-865c-572da28ab46c" containerName="mariadb-account-create-update" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.031405 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8c2151f-9cdb-445e-865c-572da28ab46c" containerName="mariadb-account-create-update" Dec 03 16:23:34 crc kubenswrapper[4998]: E1203 16:23:34.031463 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00a5807c-76d9-4a03-b8f1-1c8823915d32" containerName="mariadb-database-create" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.031518 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="00a5807c-76d9-4a03-b8f1-1c8823915d32" containerName="mariadb-database-create" Dec 03 16:23:34 crc kubenswrapper[4998]: E1203 16:23:34.031581 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20531474-cb0d-4ce8-90fc-e9974acc6d1a" containerName="mariadb-database-create" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.031637 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="20531474-cb0d-4ce8-90fc-e9974acc6d1a" containerName="mariadb-database-create" Dec 03 16:23:34 crc kubenswrapper[4998]: E1203 16:23:34.031696 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fecdd245-3377-4b65-b04d-50da92ee757b" containerName="mariadb-account-create-update" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.031772 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="fecdd245-3377-4b65-b04d-50da92ee757b" containerName="mariadb-account-create-update" Dec 03 16:23:34 crc kubenswrapper[4998]: E1203 16:23:34.031856 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fc53f31-60a9-4a2d-9d61-6c3263f95ca9" containerName="mariadb-account-create-update" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.031916 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fc53f31-60a9-4a2d-9d61-6c3263f95ca9" containerName="mariadb-account-create-update" Dec 03 16:23:34 crc kubenswrapper[4998]: E1203 16:23:34.031981 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23518b26-e4ce-495d-aada-91f784244125" containerName="mariadb-database-create" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.032039 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="23518b26-e4ce-495d-aada-91f784244125" containerName="mariadb-database-create" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.032248 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cb17c41-9a33-45df-b013-044e94f413b4" containerName="mariadb-account-create-update" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.032315 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8c2151f-9cdb-445e-865c-572da28ab46c" containerName="mariadb-account-create-update" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.032378 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="00a5807c-76d9-4a03-b8f1-1c8823915d32" containerName="mariadb-database-create" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.032437 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="23518b26-e4ce-495d-aada-91f784244125" containerName="mariadb-database-create" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.032503 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a4bd716-a01d-4e69-90ee-ae27b0e77c85" containerName="mariadb-database-create" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.032570 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="20531474-cb0d-4ce8-90fc-e9974acc6d1a" containerName="mariadb-database-create" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.032633 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="fecdd245-3377-4b65-b04d-50da92ee757b" containerName="mariadb-account-create-update" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.032695 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fc53f31-60a9-4a2d-9d61-6c3263f95ca9" containerName="mariadb-account-create-update" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.033376 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-2rqhm" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.035476 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.035561 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bpb5n" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.039602 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-2rqhm"] Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.163822 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftjd7\" (UniqueName: \"kubernetes.io/projected/de77cff6-7039-418c-8114-3c64ea4ce8e4-kube-api-access-ftjd7\") pod \"glance-db-sync-2rqhm\" (UID: \"de77cff6-7039-418c-8114-3c64ea4ce8e4\") " pod="openstack/glance-db-sync-2rqhm" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.164477 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de77cff6-7039-418c-8114-3c64ea4ce8e4-combined-ca-bundle\") pod \"glance-db-sync-2rqhm\" (UID: \"de77cff6-7039-418c-8114-3c64ea4ce8e4\") " pod="openstack/glance-db-sync-2rqhm" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.164869 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de77cff6-7039-418c-8114-3c64ea4ce8e4-config-data\") pod \"glance-db-sync-2rqhm\" (UID: \"de77cff6-7039-418c-8114-3c64ea4ce8e4\") " pod="openstack/glance-db-sync-2rqhm" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.165053 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/de77cff6-7039-418c-8114-3c64ea4ce8e4-db-sync-config-data\") pod \"glance-db-sync-2rqhm\" (UID: \"de77cff6-7039-418c-8114-3c64ea4ce8e4\") " pod="openstack/glance-db-sync-2rqhm" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.266906 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de77cff6-7039-418c-8114-3c64ea4ce8e4-combined-ca-bundle\") pod \"glance-db-sync-2rqhm\" (UID: \"de77cff6-7039-418c-8114-3c64ea4ce8e4\") " pod="openstack/glance-db-sync-2rqhm" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.267054 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de77cff6-7039-418c-8114-3c64ea4ce8e4-config-data\") pod \"glance-db-sync-2rqhm\" (UID: \"de77cff6-7039-418c-8114-3c64ea4ce8e4\") " pod="openstack/glance-db-sync-2rqhm" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.267126 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/de77cff6-7039-418c-8114-3c64ea4ce8e4-db-sync-config-data\") pod \"glance-db-sync-2rqhm\" (UID: \"de77cff6-7039-418c-8114-3c64ea4ce8e4\") " pod="openstack/glance-db-sync-2rqhm" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.267202 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftjd7\" (UniqueName: \"kubernetes.io/projected/de77cff6-7039-418c-8114-3c64ea4ce8e4-kube-api-access-ftjd7\") pod \"glance-db-sync-2rqhm\" (UID: \"de77cff6-7039-418c-8114-3c64ea4ce8e4\") " pod="openstack/glance-db-sync-2rqhm" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.277940 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de77cff6-7039-418c-8114-3c64ea4ce8e4-combined-ca-bundle\") pod \"glance-db-sync-2rqhm\" (UID: \"de77cff6-7039-418c-8114-3c64ea4ce8e4\") " pod="openstack/glance-db-sync-2rqhm" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.284363 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de77cff6-7039-418c-8114-3c64ea4ce8e4-config-data\") pod \"glance-db-sync-2rqhm\" (UID: \"de77cff6-7039-418c-8114-3c64ea4ce8e4\") " pod="openstack/glance-db-sync-2rqhm" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.296490 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/de77cff6-7039-418c-8114-3c64ea4ce8e4-db-sync-config-data\") pod \"glance-db-sync-2rqhm\" (UID: \"de77cff6-7039-418c-8114-3c64ea4ce8e4\") " pod="openstack/glance-db-sync-2rqhm" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.306333 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftjd7\" (UniqueName: \"kubernetes.io/projected/de77cff6-7039-418c-8114-3c64ea4ce8e4-kube-api-access-ftjd7\") pod \"glance-db-sync-2rqhm\" (UID: \"de77cff6-7039-418c-8114-3c64ea4ce8e4\") " pod="openstack/glance-db-sync-2rqhm" Dec 03 16:23:34 crc kubenswrapper[4998]: I1203 16:23:34.359833 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-2rqhm" Dec 03 16:23:35 crc kubenswrapper[4998]: I1203 16:23:35.342625 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" Dec 03 16:23:35 crc kubenswrapper[4998]: I1203 16:23:35.486483 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e266f963-33c9-48dc-bb10-70bd5c700d69-ovsdbserver-nb\") pod \"e266f963-33c9-48dc-bb10-70bd5c700d69\" (UID: \"e266f963-33c9-48dc-bb10-70bd5c700d69\") " Dec 03 16:23:35 crc kubenswrapper[4998]: I1203 16:23:35.486544 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnvj4\" (UniqueName: \"kubernetes.io/projected/e266f963-33c9-48dc-bb10-70bd5c700d69-kube-api-access-xnvj4\") pod \"e266f963-33c9-48dc-bb10-70bd5c700d69\" (UID: \"e266f963-33c9-48dc-bb10-70bd5c700d69\") " Dec 03 16:23:35 crc kubenswrapper[4998]: I1203 16:23:35.486673 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e266f963-33c9-48dc-bb10-70bd5c700d69-config\") pod \"e266f963-33c9-48dc-bb10-70bd5c700d69\" (UID: \"e266f963-33c9-48dc-bb10-70bd5c700d69\") " Dec 03 16:23:35 crc kubenswrapper[4998]: I1203 16:23:35.486781 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e266f963-33c9-48dc-bb10-70bd5c700d69-dns-svc\") pod \"e266f963-33c9-48dc-bb10-70bd5c700d69\" (UID: \"e266f963-33c9-48dc-bb10-70bd5c700d69\") " Dec 03 16:23:35 crc kubenswrapper[4998]: I1203 16:23:35.486832 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e266f963-33c9-48dc-bb10-70bd5c700d69-ovsdbserver-sb\") pod \"e266f963-33c9-48dc-bb10-70bd5c700d69\" (UID: \"e266f963-33c9-48dc-bb10-70bd5c700d69\") " Dec 03 16:23:35 crc kubenswrapper[4998]: I1203 16:23:35.491857 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e266f963-33c9-48dc-bb10-70bd5c700d69-kube-api-access-xnvj4" (OuterVolumeSpecName: "kube-api-access-xnvj4") pod "e266f963-33c9-48dc-bb10-70bd5c700d69" (UID: "e266f963-33c9-48dc-bb10-70bd5c700d69"). InnerVolumeSpecName "kube-api-access-xnvj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:23:35 crc kubenswrapper[4998]: I1203 16:23:35.540887 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e266f963-33c9-48dc-bb10-70bd5c700d69-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e266f963-33c9-48dc-bb10-70bd5c700d69" (UID: "e266f963-33c9-48dc-bb10-70bd5c700d69"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:35 crc kubenswrapper[4998]: I1203 16:23:35.546305 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e266f963-33c9-48dc-bb10-70bd5c700d69-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e266f963-33c9-48dc-bb10-70bd5c700d69" (UID: "e266f963-33c9-48dc-bb10-70bd5c700d69"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:35 crc kubenswrapper[4998]: I1203 16:23:35.547665 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e266f963-33c9-48dc-bb10-70bd5c700d69-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e266f963-33c9-48dc-bb10-70bd5c700d69" (UID: "e266f963-33c9-48dc-bb10-70bd5c700d69"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:35 crc kubenswrapper[4998]: I1203 16:23:35.548009 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e266f963-33c9-48dc-bb10-70bd5c700d69-config" (OuterVolumeSpecName: "config") pod "e266f963-33c9-48dc-bb10-70bd5c700d69" (UID: "e266f963-33c9-48dc-bb10-70bd5c700d69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:35 crc kubenswrapper[4998]: I1203 16:23:35.588639 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e266f963-33c9-48dc-bb10-70bd5c700d69-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:35 crc kubenswrapper[4998]: I1203 16:23:35.588685 4998 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e266f963-33c9-48dc-bb10-70bd5c700d69-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:35 crc kubenswrapper[4998]: I1203 16:23:35.588702 4998 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e266f963-33c9-48dc-bb10-70bd5c700d69-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:35 crc kubenswrapper[4998]: I1203 16:23:35.588714 4998 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e266f963-33c9-48dc-bb10-70bd5c700d69-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:35 crc kubenswrapper[4998]: I1203 16:23:35.588727 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnvj4\" (UniqueName: \"kubernetes.io/projected/e266f963-33c9-48dc-bb10-70bd5c700d69-kube-api-access-xnvj4\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:35 crc kubenswrapper[4998]: I1203 16:23:35.641037 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" event={"ID":"e266f963-33c9-48dc-bb10-70bd5c700d69","Type":"ContainerDied","Data":"7e6b8fcb2c28749e83376531d1ca7c381481f59b2a757cb53d68873facac9224"} Dec 03 16:23:35 crc kubenswrapper[4998]: I1203 16:23:35.641097 4998 scope.go:117] "RemoveContainer" containerID="8a02ff22240d153f314a59c60dd4559c01cd622cfce7f13356e6514f03fde803" Dec 03 16:23:35 crc kubenswrapper[4998]: I1203 16:23:35.641250 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" Dec 03 16:23:35 crc kubenswrapper[4998]: I1203 16:23:35.688689 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55ddfd5dfc-9clrt"] Dec 03 16:23:35 crc kubenswrapper[4998]: I1203 16:23:35.688725 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55ddfd5dfc-9clrt"] Dec 03 16:23:36 crc kubenswrapper[4998]: I1203 16:23:36.217141 4998 scope.go:117] "RemoveContainer" containerID="241d5701046d2695d1096b4ebf57a93394e22c261fa3f23621e2e3b2ec074ffc" Dec 03 16:23:36 crc kubenswrapper[4998]: I1203 16:23:36.648814 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a70537b1-93bc-48c3-9bef-fd9030158cad","Type":"ContainerStarted","Data":"a7bc207f9aabc3bd16ffd9cad967412d25c52c69865f816487cd0b58b6c4d55b"} Dec 03 16:23:36 crc kubenswrapper[4998]: I1203 16:23:36.865309 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-2rqhm"] Dec 03 16:23:37 crc kubenswrapper[4998]: I1203 16:23:37.136657 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-55ddfd5dfc-9clrt" podUID="e266f963-33c9-48dc-bb10-70bd5c700d69" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.121:5353: i/o timeout" Dec 03 16:23:37 crc kubenswrapper[4998]: I1203 16:23:37.662238 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-rjhmp" event={"ID":"edc74770-171e-4ade-8dcd-bcaead186741","Type":"ContainerStarted","Data":"4140050d1ad6858044ba4f30d05c0f46fdc7befefd812effe581a96fa4ad4b18"} Dec 03 16:23:37 crc kubenswrapper[4998]: I1203 16:23:37.664673 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-nj4w6" event={"ID":"e11bf44e-799b-472b-aafb-89d5d071ad91","Type":"ContainerStarted","Data":"4f14427fc6d32bd017613ec2b7d3009f7f7eab8658a1964b632c6af7dbc42fde"} Dec 03 16:23:37 crc kubenswrapper[4998]: I1203 16:23:37.667374 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-2rqhm" event={"ID":"de77cff6-7039-418c-8114-3c64ea4ce8e4","Type":"ContainerStarted","Data":"49b4dd73ac30d196a03ffee484d4181b02cb8c15f2de3aceb1023b7abb6709a4"} Dec 03 16:23:37 crc kubenswrapper[4998]: I1203 16:23:37.681912 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-db-sync-rjhmp" podStartSLOduration=3.166120559 podStartE2EDuration="14.681887679s" podCreationTimestamp="2025-12-03 16:23:23 +0000 UTC" firstStartedPulling="2025-12-03 16:23:24.804169928 +0000 UTC m=+1183.415870151" lastFinishedPulling="2025-12-03 16:23:36.319937048 +0000 UTC m=+1194.931637271" observedRunningTime="2025-12-03 16:23:37.679675833 +0000 UTC m=+1196.291376056" watchObservedRunningTime="2025-12-03 16:23:37.681887679 +0000 UTC m=+1196.293587922" Dec 03 16:23:37 crc kubenswrapper[4998]: I1203 16:23:37.692133 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e266f963-33c9-48dc-bb10-70bd5c700d69" path="/var/lib/kubelet/pods/e266f963-33c9-48dc-bb10-70bd5c700d69/volumes" Dec 03 16:23:37 crc kubenswrapper[4998]: I1203 16:23:37.703985 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-nj4w6" podStartSLOduration=2.668907538 podStartE2EDuration="15.703964905s" podCreationTimestamp="2025-12-03 16:23:22 +0000 UTC" firstStartedPulling="2025-12-03 16:23:23.182588134 +0000 UTC m=+1181.794288357" lastFinishedPulling="2025-12-03 16:23:36.217645501 +0000 UTC m=+1194.829345724" observedRunningTime="2025-12-03 16:23:37.694786334 +0000 UTC m=+1196.306486577" watchObservedRunningTime="2025-12-03 16:23:37.703964905 +0000 UTC m=+1196.315665148" Dec 03 16:23:39 crc kubenswrapper[4998]: I1203 16:23:39.693008 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a70537b1-93bc-48c3-9bef-fd9030158cad","Type":"ContainerStarted","Data":"d6a4eaeddd7b46a9b6c18edbb3bab49896a9ae66616e1863c624492c605f1345"} Dec 03 16:23:39 crc kubenswrapper[4998]: I1203 16:23:39.693538 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a70537b1-93bc-48c3-9bef-fd9030158cad","Type":"ContainerStarted","Data":"21be89612a3b0e4cd79f2ace91e7ce2132daa6eb2b02c74b17c740d1d544d851"} Dec 03 16:23:39 crc kubenswrapper[4998]: I1203 16:23:39.722719 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=23.722702774 podStartE2EDuration="23.722702774s" podCreationTimestamp="2025-12-03 16:23:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:23:39.715459072 +0000 UTC m=+1198.327159295" watchObservedRunningTime="2025-12-03 16:23:39.722702774 +0000 UTC m=+1198.334402997" Dec 03 16:23:40 crc kubenswrapper[4998]: I1203 16:23:40.706212 4998 generic.go:334] "Generic (PLEG): container finished" podID="edc74770-171e-4ade-8dcd-bcaead186741" containerID="4140050d1ad6858044ba4f30d05c0f46fdc7befefd812effe581a96fa4ad4b18" exitCode=0 Dec 03 16:23:40 crc kubenswrapper[4998]: I1203 16:23:40.707336 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-rjhmp" event={"ID":"edc74770-171e-4ade-8dcd-bcaead186741","Type":"ContainerDied","Data":"4140050d1ad6858044ba4f30d05c0f46fdc7befefd812effe581a96fa4ad4b18"} Dec 03 16:23:41 crc kubenswrapper[4998]: I1203 16:23:41.738560 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:42 crc kubenswrapper[4998]: I1203 16:23:42.747310 4998 generic.go:334] "Generic (PLEG): container finished" podID="e11bf44e-799b-472b-aafb-89d5d071ad91" containerID="4f14427fc6d32bd017613ec2b7d3009f7f7eab8658a1964b632c6af7dbc42fde" exitCode=0 Dec 03 16:23:42 crc kubenswrapper[4998]: I1203 16:23:42.747385 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-nj4w6" event={"ID":"e11bf44e-799b-472b-aafb-89d5d071ad91","Type":"ContainerDied","Data":"4f14427fc6d32bd017613ec2b7d3009f7f7eab8658a1964b632c6af7dbc42fde"} Dec 03 16:23:46 crc kubenswrapper[4998]: I1203 16:23:46.737671 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:46 crc kubenswrapper[4998]: I1203 16:23:46.744363 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:46 crc kubenswrapper[4998]: I1203 16:23:46.794405 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 03 16:23:55 crc kubenswrapper[4998]: E1203 16:23:55.538148 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.32:5001/podified-master-centos10/openstack-glance-api:watcher_latest" Dec 03 16:23:55 crc kubenswrapper[4998]: E1203 16:23:55.538743 4998 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.32:5001/podified-master-centos10/openstack-glance-api:watcher_latest" Dec 03 16:23:55 crc kubenswrapper[4998]: E1203 16:23:55.538892 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:38.102.83.32:5001/podified-master-centos10/openstack-glance-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ftjd7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-2rqhm_openstack(de77cff6-7039-418c-8114-3c64ea4ce8e4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 16:23:55 crc kubenswrapper[4998]: E1203 16:23:55.540791 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-2rqhm" podUID="de77cff6-7039-418c-8114-3c64ea4ce8e4" Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.561884 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-rjhmp" Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.581098 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-nj4w6" Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.679614 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/edc74770-171e-4ade-8dcd-bcaead186741-db-sync-config-data\") pod \"edc74770-171e-4ade-8dcd-bcaead186741\" (UID: \"edc74770-171e-4ade-8dcd-bcaead186741\") " Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.679702 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xmxh\" (UniqueName: \"kubernetes.io/projected/edc74770-171e-4ade-8dcd-bcaead186741-kube-api-access-4xmxh\") pod \"edc74770-171e-4ade-8dcd-bcaead186741\" (UID: \"edc74770-171e-4ade-8dcd-bcaead186741\") " Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.679736 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edc74770-171e-4ade-8dcd-bcaead186741-config-data\") pod \"edc74770-171e-4ade-8dcd-bcaead186741\" (UID: \"edc74770-171e-4ade-8dcd-bcaead186741\") " Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.679787 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rx6kd\" (UniqueName: \"kubernetes.io/projected/e11bf44e-799b-472b-aafb-89d5d071ad91-kube-api-access-rx6kd\") pod \"e11bf44e-799b-472b-aafb-89d5d071ad91\" (UID: \"e11bf44e-799b-472b-aafb-89d5d071ad91\") " Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.679885 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edc74770-171e-4ade-8dcd-bcaead186741-combined-ca-bundle\") pod \"edc74770-171e-4ade-8dcd-bcaead186741\" (UID: \"edc74770-171e-4ade-8dcd-bcaead186741\") " Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.679991 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e11bf44e-799b-472b-aafb-89d5d071ad91-config-data\") pod \"e11bf44e-799b-472b-aafb-89d5d071ad91\" (UID: \"e11bf44e-799b-472b-aafb-89d5d071ad91\") " Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.680011 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e11bf44e-799b-472b-aafb-89d5d071ad91-combined-ca-bundle\") pod \"e11bf44e-799b-472b-aafb-89d5d071ad91\" (UID: \"e11bf44e-799b-472b-aafb-89d5d071ad91\") " Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.713330 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e11bf44e-799b-472b-aafb-89d5d071ad91-kube-api-access-rx6kd" (OuterVolumeSpecName: "kube-api-access-rx6kd") pod "e11bf44e-799b-472b-aafb-89d5d071ad91" (UID: "e11bf44e-799b-472b-aafb-89d5d071ad91"). InnerVolumeSpecName "kube-api-access-rx6kd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.715392 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edc74770-171e-4ade-8dcd-bcaead186741-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "edc74770-171e-4ade-8dcd-bcaead186741" (UID: "edc74770-171e-4ade-8dcd-bcaead186741"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.737445 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edc74770-171e-4ade-8dcd-bcaead186741-kube-api-access-4xmxh" (OuterVolumeSpecName: "kube-api-access-4xmxh") pod "edc74770-171e-4ade-8dcd-bcaead186741" (UID: "edc74770-171e-4ade-8dcd-bcaead186741"). InnerVolumeSpecName "kube-api-access-4xmxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.757677 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edc74770-171e-4ade-8dcd-bcaead186741-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "edc74770-171e-4ade-8dcd-bcaead186741" (UID: "edc74770-171e-4ade-8dcd-bcaead186741"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.782276 4998 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/edc74770-171e-4ade-8dcd-bcaead186741-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.782521 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xmxh\" (UniqueName: \"kubernetes.io/projected/edc74770-171e-4ade-8dcd-bcaead186741-kube-api-access-4xmxh\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.782601 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rx6kd\" (UniqueName: \"kubernetes.io/projected/e11bf44e-799b-472b-aafb-89d5d071ad91-kube-api-access-rx6kd\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.782669 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edc74770-171e-4ade-8dcd-bcaead186741-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.803549 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e11bf44e-799b-472b-aafb-89d5d071ad91-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e11bf44e-799b-472b-aafb-89d5d071ad91" (UID: "e11bf44e-799b-472b-aafb-89d5d071ad91"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.825630 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e11bf44e-799b-472b-aafb-89d5d071ad91-config-data" (OuterVolumeSpecName: "config-data") pod "e11bf44e-799b-472b-aafb-89d5d071ad91" (UID: "e11bf44e-799b-472b-aafb-89d5d071ad91"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.834362 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edc74770-171e-4ade-8dcd-bcaead186741-config-data" (OuterVolumeSpecName: "config-data") pod "edc74770-171e-4ade-8dcd-bcaead186741" (UID: "edc74770-171e-4ade-8dcd-bcaead186741"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.869911 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-nj4w6" Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.869892 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-nj4w6" event={"ID":"e11bf44e-799b-472b-aafb-89d5d071ad91","Type":"ContainerDied","Data":"486591e7c40536756dccef5049657a3f0842bd5227302a7c6351ebcf1195dbea"} Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.870256 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="486591e7c40536756dccef5049657a3f0842bd5227302a7c6351ebcf1195dbea" Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.873254 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-rjhmp" Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.873681 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-rjhmp" event={"ID":"edc74770-171e-4ade-8dcd-bcaead186741","Type":"ContainerDied","Data":"e001f3a8a1b54ab36a1848a785335e49a5222831bbedf30382e2e8ae6f6ec6b7"} Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.873707 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e001f3a8a1b54ab36a1848a785335e49a5222831bbedf30382e2e8ae6f6ec6b7" Dec 03 16:23:55 crc kubenswrapper[4998]: E1203 16:23:55.874109 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.32:5001/podified-master-centos10/openstack-glance-api:watcher_latest\\\"\"" pod="openstack/glance-db-sync-2rqhm" podUID="de77cff6-7039-418c-8114-3c64ea4ce8e4" Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.883747 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e11bf44e-799b-472b-aafb-89d5d071ad91-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.883790 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e11bf44e-799b-472b-aafb-89d5d071ad91-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:55 crc kubenswrapper[4998]: I1203 16:23:55.883802 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edc74770-171e-4ade-8dcd-bcaead186741-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:56 crc kubenswrapper[4998]: I1203 16:23:56.969495 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-hd2nb"] Dec 03 16:23:56 crc kubenswrapper[4998]: E1203 16:23:56.970105 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e11bf44e-799b-472b-aafb-89d5d071ad91" containerName="keystone-db-sync" Dec 03 16:23:56 crc kubenswrapper[4998]: I1203 16:23:56.970118 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="e11bf44e-799b-472b-aafb-89d5d071ad91" containerName="keystone-db-sync" Dec 03 16:23:56 crc kubenswrapper[4998]: E1203 16:23:56.970137 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e266f963-33c9-48dc-bb10-70bd5c700d69" containerName="init" Dec 03 16:23:56 crc kubenswrapper[4998]: I1203 16:23:56.970143 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="e266f963-33c9-48dc-bb10-70bd5c700d69" containerName="init" Dec 03 16:23:56 crc kubenswrapper[4998]: E1203 16:23:56.970167 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e266f963-33c9-48dc-bb10-70bd5c700d69" containerName="dnsmasq-dns" Dec 03 16:23:56 crc kubenswrapper[4998]: I1203 16:23:56.970173 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="e266f963-33c9-48dc-bb10-70bd5c700d69" containerName="dnsmasq-dns" Dec 03 16:23:56 crc kubenswrapper[4998]: E1203 16:23:56.970185 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edc74770-171e-4ade-8dcd-bcaead186741" containerName="watcher-db-sync" Dec 03 16:23:56 crc kubenswrapper[4998]: I1203 16:23:56.970190 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="edc74770-171e-4ade-8dcd-bcaead186741" containerName="watcher-db-sync" Dec 03 16:23:56 crc kubenswrapper[4998]: I1203 16:23:56.970340 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="e266f963-33c9-48dc-bb10-70bd5c700d69" containerName="dnsmasq-dns" Dec 03 16:23:56 crc kubenswrapper[4998]: I1203 16:23:56.970362 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="edc74770-171e-4ade-8dcd-bcaead186741" containerName="watcher-db-sync" Dec 03 16:23:56 crc kubenswrapper[4998]: I1203 16:23:56.970372 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="e11bf44e-799b-472b-aafb-89d5d071ad91" containerName="keystone-db-sync" Dec 03 16:23:56 crc kubenswrapper[4998]: I1203 16:23:56.971032 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hd2nb" Dec 03 16:23:56 crc kubenswrapper[4998]: I1203 16:23:56.975874 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 16:23:56 crc kubenswrapper[4998]: I1203 16:23:56.976012 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 16:23:56 crc kubenswrapper[4998]: I1203 16:23:56.976031 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 16:23:56 crc kubenswrapper[4998]: I1203 16:23:56.976143 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ggfph" Dec 03 16:23:56 crc kubenswrapper[4998]: I1203 16:23:56.976200 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.020823 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hd2nb"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.037909 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ff6fd6c5-fnwm8"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.039644 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ff6fd6c5-fnwm8" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.044820 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ff6fd6c5-fnwm8"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.066841 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.068462 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.086790 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.086996 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-watcher-dockercfg-vlg9j" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.107852 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.120619 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-applier-0"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.121680 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.128062 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-applier-config-data" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.138208 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.149684 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e968cb62-c964-46f9-ac6d-6cb674c8bf22-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"e968cb62-c964-46f9-ac6d-6cb674c8bf22\") " pod="openstack/watcher-api-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.149726 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e968cb62-c964-46f9-ac6d-6cb674c8bf22-config-data\") pod \"watcher-api-0\" (UID: \"e968cb62-c964-46f9-ac6d-6cb674c8bf22\") " pod="openstack/watcher-api-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.149820 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-dns-svc\") pod \"dnsmasq-dns-5ff6fd6c5-fnwm8\" (UID: \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\") " pod="openstack/dnsmasq-dns-5ff6fd6c5-fnwm8" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.149842 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/e968cb62-c964-46f9-ac6d-6cb674c8bf22-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"e968cb62-c964-46f9-ac6d-6cb674c8bf22\") " pod="openstack/watcher-api-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.149863 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-credential-keys\") pod \"keystone-bootstrap-hd2nb\" (UID: \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\") " pod="openstack/keystone-bootstrap-hd2nb" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.149883 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbz7r\" (UniqueName: \"kubernetes.io/projected/e968cb62-c964-46f9-ac6d-6cb674c8bf22-kube-api-access-rbz7r\") pod \"watcher-api-0\" (UID: \"e968cb62-c964-46f9-ac6d-6cb674c8bf22\") " pod="openstack/watcher-api-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.149917 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e968cb62-c964-46f9-ac6d-6cb674c8bf22-logs\") pod \"watcher-api-0\" (UID: \"e968cb62-c964-46f9-ac6d-6cb674c8bf22\") " pod="openstack/watcher-api-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.149944 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-fernet-keys\") pod \"keystone-bootstrap-hd2nb\" (UID: \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\") " pod="openstack/keystone-bootstrap-hd2nb" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.149960 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-dns-swift-storage-0\") pod \"dnsmasq-dns-5ff6fd6c5-fnwm8\" (UID: \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\") " pod="openstack/dnsmasq-dns-5ff6fd6c5-fnwm8" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.149978 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjmsl\" (UniqueName: \"kubernetes.io/projected/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-kube-api-access-gjmsl\") pod \"keystone-bootstrap-hd2nb\" (UID: \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\") " pod="openstack/keystone-bootstrap-hd2nb" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.149992 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-scripts\") pod \"keystone-bootstrap-hd2nb\" (UID: \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\") " pod="openstack/keystone-bootstrap-hd2nb" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.150008 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj2x6\" (UniqueName: \"kubernetes.io/projected/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-kube-api-access-dj2x6\") pod \"dnsmasq-dns-5ff6fd6c5-fnwm8\" (UID: \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\") " pod="openstack/dnsmasq-dns-5ff6fd6c5-fnwm8" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.150043 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-config-data\") pod \"keystone-bootstrap-hd2nb\" (UID: \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\") " pod="openstack/keystone-bootstrap-hd2nb" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.150058 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-config\") pod \"dnsmasq-dns-5ff6fd6c5-fnwm8\" (UID: \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\") " pod="openstack/dnsmasq-dns-5ff6fd6c5-fnwm8" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.150082 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-ovsdbserver-sb\") pod \"dnsmasq-dns-5ff6fd6c5-fnwm8\" (UID: \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\") " pod="openstack/dnsmasq-dns-5ff6fd6c5-fnwm8" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.150108 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-combined-ca-bundle\") pod \"keystone-bootstrap-hd2nb\" (UID: \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\") " pod="openstack/keystone-bootstrap-hd2nb" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.150137 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-ovsdbserver-nb\") pod \"dnsmasq-dns-5ff6fd6c5-fnwm8\" (UID: \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\") " pod="openstack/dnsmasq-dns-5ff6fd6c5-fnwm8" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.165734 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-8lz9j"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.166909 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8lz9j" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.174662 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.174871 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.174975 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-8dr6s" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.184648 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-8lz9j"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.233172 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-648fb89557-mvw59"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.234566 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-648fb89557-mvw59" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.238705 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.240613 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.240821 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-pt2l7" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.241445 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.252555 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-credential-keys\") pod \"keystone-bootstrap-hd2nb\" (UID: \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\") " pod="openstack/keystone-bootstrap-hd2nb" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.252608 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbz7r\" (UniqueName: \"kubernetes.io/projected/e968cb62-c964-46f9-ac6d-6cb674c8bf22-kube-api-access-rbz7r\") pod \"watcher-api-0\" (UID: \"e968cb62-c964-46f9-ac6d-6cb674c8bf22\") " pod="openstack/watcher-api-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.252652 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fcmz\" (UniqueName: \"kubernetes.io/projected/5b94f565-0948-4e4f-9636-2833d43bb829-kube-api-access-8fcmz\") pod \"watcher-applier-0\" (UID: \"5b94f565-0948-4e4f-9636-2833d43bb829\") " pod="openstack/watcher-applier-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.252671 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82aa9d23-457c-45a0-9729-e5a803738237-config-data\") pod \"cinder-db-sync-8lz9j\" (UID: \"82aa9d23-457c-45a0-9729-e5a803738237\") " pod="openstack/cinder-db-sync-8lz9j" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.252697 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e968cb62-c964-46f9-ac6d-6cb674c8bf22-logs\") pod \"watcher-api-0\" (UID: \"e968cb62-c964-46f9-ac6d-6cb674c8bf22\") " pod="openstack/watcher-api-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.252713 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82aa9d23-457c-45a0-9729-e5a803738237-combined-ca-bundle\") pod \"cinder-db-sync-8lz9j\" (UID: \"82aa9d23-457c-45a0-9729-e5a803738237\") " pod="openstack/cinder-db-sync-8lz9j" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.252765 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-fernet-keys\") pod \"keystone-bootstrap-hd2nb\" (UID: \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\") " pod="openstack/keystone-bootstrap-hd2nb" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.252783 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-dns-swift-storage-0\") pod \"dnsmasq-dns-5ff6fd6c5-fnwm8\" (UID: \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\") " pod="openstack/dnsmasq-dns-5ff6fd6c5-fnwm8" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.252799 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82aa9d23-457c-45a0-9729-e5a803738237-scripts\") pod \"cinder-db-sync-8lz9j\" (UID: \"82aa9d23-457c-45a0-9729-e5a803738237\") " pod="openstack/cinder-db-sync-8lz9j" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.252818 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjmsl\" (UniqueName: \"kubernetes.io/projected/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-kube-api-access-gjmsl\") pod \"keystone-bootstrap-hd2nb\" (UID: \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\") " pod="openstack/keystone-bootstrap-hd2nb" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.252837 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-scripts\") pod \"keystone-bootstrap-hd2nb\" (UID: \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\") " pod="openstack/keystone-bootstrap-hd2nb" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.252855 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj2x6\" (UniqueName: \"kubernetes.io/projected/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-kube-api-access-dj2x6\") pod \"dnsmasq-dns-5ff6fd6c5-fnwm8\" (UID: \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\") " pod="openstack/dnsmasq-dns-5ff6fd6c5-fnwm8" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.252885 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b94f565-0948-4e4f-9636-2833d43bb829-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"5b94f565-0948-4e4f-9636-2833d43bb829\") " pod="openstack/watcher-applier-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.252916 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-config-data\") pod \"keystone-bootstrap-hd2nb\" (UID: \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\") " pod="openstack/keystone-bootstrap-hd2nb" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.252932 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82aa9d23-457c-45a0-9729-e5a803738237-etc-machine-id\") pod \"cinder-db-sync-8lz9j\" (UID: \"82aa9d23-457c-45a0-9729-e5a803738237\") " pod="openstack/cinder-db-sync-8lz9j" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.252948 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-config\") pod \"dnsmasq-dns-5ff6fd6c5-fnwm8\" (UID: \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\") " pod="openstack/dnsmasq-dns-5ff6fd6c5-fnwm8" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.252983 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/82aa9d23-457c-45a0-9729-e5a803738237-db-sync-config-data\") pod \"cinder-db-sync-8lz9j\" (UID: \"82aa9d23-457c-45a0-9729-e5a803738237\") " pod="openstack/cinder-db-sync-8lz9j" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.253001 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnlxq\" (UniqueName: \"kubernetes.io/projected/82aa9d23-457c-45a0-9729-e5a803738237-kube-api-access-jnlxq\") pod \"cinder-db-sync-8lz9j\" (UID: \"82aa9d23-457c-45a0-9729-e5a803738237\") " pod="openstack/cinder-db-sync-8lz9j" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.253029 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-ovsdbserver-sb\") pod \"dnsmasq-dns-5ff6fd6c5-fnwm8\" (UID: \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\") " pod="openstack/dnsmasq-dns-5ff6fd6c5-fnwm8" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.253060 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-combined-ca-bundle\") pod \"keystone-bootstrap-hd2nb\" (UID: \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\") " pod="openstack/keystone-bootstrap-hd2nb" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.253101 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-ovsdbserver-nb\") pod \"dnsmasq-dns-5ff6fd6c5-fnwm8\" (UID: \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\") " pod="openstack/dnsmasq-dns-5ff6fd6c5-fnwm8" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.253119 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b94f565-0948-4e4f-9636-2833d43bb829-logs\") pod \"watcher-applier-0\" (UID: \"5b94f565-0948-4e4f-9636-2833d43bb829\") " pod="openstack/watcher-applier-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.253147 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e968cb62-c964-46f9-ac6d-6cb674c8bf22-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"e968cb62-c964-46f9-ac6d-6cb674c8bf22\") " pod="openstack/watcher-api-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.253163 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e968cb62-c964-46f9-ac6d-6cb674c8bf22-config-data\") pod \"watcher-api-0\" (UID: \"e968cb62-c964-46f9-ac6d-6cb674c8bf22\") " pod="openstack/watcher-api-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.253185 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-dns-svc\") pod \"dnsmasq-dns-5ff6fd6c5-fnwm8\" (UID: \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\") " pod="openstack/dnsmasq-dns-5ff6fd6c5-fnwm8" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.253208 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/e968cb62-c964-46f9-ac6d-6cb674c8bf22-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"e968cb62-c964-46f9-ac6d-6cb674c8bf22\") " pod="openstack/watcher-api-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.253225 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b94f565-0948-4e4f-9636-2833d43bb829-config-data\") pod \"watcher-applier-0\" (UID: \"5b94f565-0948-4e4f-9636-2833d43bb829\") " pod="openstack/watcher-applier-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.253934 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e968cb62-c964-46f9-ac6d-6cb674c8bf22-logs\") pod \"watcher-api-0\" (UID: \"e968cb62-c964-46f9-ac6d-6cb674c8bf22\") " pod="openstack/watcher-api-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.254465 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-648fb89557-mvw59"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.255347 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-config\") pod \"dnsmasq-dns-5ff6fd6c5-fnwm8\" (UID: \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\") " pod="openstack/dnsmasq-dns-5ff6fd6c5-fnwm8" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.260496 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-dns-swift-storage-0\") pod \"dnsmasq-dns-5ff6fd6c5-fnwm8\" (UID: \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\") " pod="openstack/dnsmasq-dns-5ff6fd6c5-fnwm8" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.261621 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-ovsdbserver-sb\") pod \"dnsmasq-dns-5ff6fd6c5-fnwm8\" (UID: \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\") " pod="openstack/dnsmasq-dns-5ff6fd6c5-fnwm8" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.266001 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-ovsdbserver-nb\") pod \"dnsmasq-dns-5ff6fd6c5-fnwm8\" (UID: \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\") " pod="openstack/dnsmasq-dns-5ff6fd6c5-fnwm8" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.267824 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-dns-svc\") pod \"dnsmasq-dns-5ff6fd6c5-fnwm8\" (UID: \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\") " pod="openstack/dnsmasq-dns-5ff6fd6c5-fnwm8" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.272998 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e968cb62-c964-46f9-ac6d-6cb674c8bf22-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"e968cb62-c964-46f9-ac6d-6cb674c8bf22\") " pod="openstack/watcher-api-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.275134 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-scripts\") pod \"keystone-bootstrap-hd2nb\" (UID: \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\") " pod="openstack/keystone-bootstrap-hd2nb" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.277249 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/e968cb62-c964-46f9-ac6d-6cb674c8bf22-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"e968cb62-c964-46f9-ac6d-6cb674c8bf22\") " pod="openstack/watcher-api-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.279694 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e968cb62-c964-46f9-ac6d-6cb674c8bf22-config-data\") pod \"watcher-api-0\" (UID: \"e968cb62-c964-46f9-ac6d-6cb674c8bf22\") " pod="openstack/watcher-api-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.284073 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-fernet-keys\") pod \"keystone-bootstrap-hd2nb\" (UID: \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\") " pod="openstack/keystone-bootstrap-hd2nb" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.298495 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj2x6\" (UniqueName: \"kubernetes.io/projected/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-kube-api-access-dj2x6\") pod \"dnsmasq-dns-5ff6fd6c5-fnwm8\" (UID: \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\") " pod="openstack/dnsmasq-dns-5ff6fd6c5-fnwm8" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.303359 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-combined-ca-bundle\") pod \"keystone-bootstrap-hd2nb\" (UID: \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\") " pod="openstack/keystone-bootstrap-hd2nb" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.304332 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbz7r\" (UniqueName: \"kubernetes.io/projected/e968cb62-c964-46f9-ac6d-6cb674c8bf22-kube-api-access-rbz7r\") pod \"watcher-api-0\" (UID: \"e968cb62-c964-46f9-ac6d-6cb674c8bf22\") " pod="openstack/watcher-api-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.309219 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjmsl\" (UniqueName: \"kubernetes.io/projected/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-kube-api-access-gjmsl\") pod \"keystone-bootstrap-hd2nb\" (UID: \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\") " pod="openstack/keystone-bootstrap-hd2nb" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.327492 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.328581 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-credential-keys\") pod \"keystone-bootstrap-hd2nb\" (UID: \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\") " pod="openstack/keystone-bootstrap-hd2nb" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.328617 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.330060 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-config-data\") pod \"keystone-bootstrap-hd2nb\" (UID: \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\") " pod="openstack/keystone-bootstrap-hd2nb" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.339068 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.356800 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x5cz\" (UniqueName: \"kubernetes.io/projected/53f87c55-ecee-4d26-8244-6181b21dece0-kube-api-access-2x5cz\") pod \"horizon-648fb89557-mvw59\" (UID: \"53f87c55-ecee-4d26-8244-6181b21dece0\") " pod="openstack/horizon-648fb89557-mvw59" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.356845 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53f87c55-ecee-4d26-8244-6181b21dece0-scripts\") pod \"horizon-648fb89557-mvw59\" (UID: \"53f87c55-ecee-4d26-8244-6181b21dece0\") " pod="openstack/horizon-648fb89557-mvw59" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.356878 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b94f565-0948-4e4f-9636-2833d43bb829-config-data\") pod \"watcher-applier-0\" (UID: \"5b94f565-0948-4e4f-9636-2833d43bb829\") " pod="openstack/watcher-applier-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.356900 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53f87c55-ecee-4d26-8244-6181b21dece0-logs\") pod \"horizon-648fb89557-mvw59\" (UID: \"53f87c55-ecee-4d26-8244-6181b21dece0\") " pod="openstack/horizon-648fb89557-mvw59" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.356931 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fcmz\" (UniqueName: \"kubernetes.io/projected/5b94f565-0948-4e4f-9636-2833d43bb829-kube-api-access-8fcmz\") pod \"watcher-applier-0\" (UID: \"5b94f565-0948-4e4f-9636-2833d43bb829\") " pod="openstack/watcher-applier-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.356947 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82aa9d23-457c-45a0-9729-e5a803738237-config-data\") pod \"cinder-db-sync-8lz9j\" (UID: \"82aa9d23-457c-45a0-9729-e5a803738237\") " pod="openstack/cinder-db-sync-8lz9j" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.356969 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82aa9d23-457c-45a0-9729-e5a803738237-combined-ca-bundle\") pod \"cinder-db-sync-8lz9j\" (UID: \"82aa9d23-457c-45a0-9729-e5a803738237\") " pod="openstack/cinder-db-sync-8lz9j" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.356988 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/53f87c55-ecee-4d26-8244-6181b21dece0-horizon-secret-key\") pod \"horizon-648fb89557-mvw59\" (UID: \"53f87c55-ecee-4d26-8244-6181b21dece0\") " pod="openstack/horizon-648fb89557-mvw59" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.357011 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82aa9d23-457c-45a0-9729-e5a803738237-scripts\") pod \"cinder-db-sync-8lz9j\" (UID: \"82aa9d23-457c-45a0-9729-e5a803738237\") " pod="openstack/cinder-db-sync-8lz9j" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.357026 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53f87c55-ecee-4d26-8244-6181b21dece0-config-data\") pod \"horizon-648fb89557-mvw59\" (UID: \"53f87c55-ecee-4d26-8244-6181b21dece0\") " pod="openstack/horizon-648fb89557-mvw59" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.357059 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b94f565-0948-4e4f-9636-2833d43bb829-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"5b94f565-0948-4e4f-9636-2833d43bb829\") " pod="openstack/watcher-applier-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.357083 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82aa9d23-457c-45a0-9729-e5a803738237-etc-machine-id\") pod \"cinder-db-sync-8lz9j\" (UID: \"82aa9d23-457c-45a0-9729-e5a803738237\") " pod="openstack/cinder-db-sync-8lz9j" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.357097 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/82aa9d23-457c-45a0-9729-e5a803738237-db-sync-config-data\") pod \"cinder-db-sync-8lz9j\" (UID: \"82aa9d23-457c-45a0-9729-e5a803738237\") " pod="openstack/cinder-db-sync-8lz9j" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.357114 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnlxq\" (UniqueName: \"kubernetes.io/projected/82aa9d23-457c-45a0-9729-e5a803738237-kube-api-access-jnlxq\") pod \"cinder-db-sync-8lz9j\" (UID: \"82aa9d23-457c-45a0-9729-e5a803738237\") " pod="openstack/cinder-db-sync-8lz9j" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.357160 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b94f565-0948-4e4f-9636-2833d43bb829-logs\") pod \"watcher-applier-0\" (UID: \"5b94f565-0948-4e4f-9636-2833d43bb829\") " pod="openstack/watcher-applier-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.357573 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b94f565-0948-4e4f-9636-2833d43bb829-logs\") pod \"watcher-applier-0\" (UID: \"5b94f565-0948-4e4f-9636-2833d43bb829\") " pod="openstack/watcher-applier-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.360839 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82aa9d23-457c-45a0-9729-e5a803738237-etc-machine-id\") pod \"cinder-db-sync-8lz9j\" (UID: \"82aa9d23-457c-45a0-9729-e5a803738237\") " pod="openstack/cinder-db-sync-8lz9j" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.367107 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.375424 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82aa9d23-457c-45a0-9729-e5a803738237-config-data\") pod \"cinder-db-sync-8lz9j\" (UID: \"82aa9d23-457c-45a0-9729-e5a803738237\") " pod="openstack/cinder-db-sync-8lz9j" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.375997 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ff6fd6c5-fnwm8" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.378723 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/82aa9d23-457c-45a0-9729-e5a803738237-db-sync-config-data\") pod \"cinder-db-sync-8lz9j\" (UID: \"82aa9d23-457c-45a0-9729-e5a803738237\") " pod="openstack/cinder-db-sync-8lz9j" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.384416 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b94f565-0948-4e4f-9636-2833d43bb829-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"5b94f565-0948-4e4f-9636-2833d43bb829\") " pod="openstack/watcher-applier-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.390956 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b94f565-0948-4e4f-9636-2833d43bb829-config-data\") pod \"watcher-applier-0\" (UID: \"5b94f565-0948-4e4f-9636-2833d43bb829\") " pod="openstack/watcher-applier-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.391309 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82aa9d23-457c-45a0-9729-e5a803738237-scripts\") pod \"cinder-db-sync-8lz9j\" (UID: \"82aa9d23-457c-45a0-9729-e5a803738237\") " pod="openstack/cinder-db-sync-8lz9j" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.397400 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82aa9d23-457c-45a0-9729-e5a803738237-combined-ca-bundle\") pod \"cinder-db-sync-8lz9j\" (UID: \"82aa9d23-457c-45a0-9729-e5a803738237\") " pod="openstack/cinder-db-sync-8lz9j" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.401866 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fcmz\" (UniqueName: \"kubernetes.io/projected/5b94f565-0948-4e4f-9636-2833d43bb829-kube-api-access-8fcmz\") pod \"watcher-applier-0\" (UID: \"5b94f565-0948-4e4f-9636-2833d43bb829\") " pod="openstack/watcher-applier-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.412445 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnlxq\" (UniqueName: \"kubernetes.io/projected/82aa9d23-457c-45a0-9729-e5a803738237-kube-api-access-jnlxq\") pod \"cinder-db-sync-8lz9j\" (UID: \"82aa9d23-457c-45a0-9729-e5a803738237\") " pod="openstack/cinder-db-sync-8lz9j" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.423106 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.445079 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-2g9pt"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.446389 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2g9pt" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.450840 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.450998 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-f986n" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.459828 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.460231 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.461052 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53f87c55-ecee-4d26-8244-6181b21dece0-config-data\") pod \"horizon-648fb89557-mvw59\" (UID: \"53f87c55-ecee-4d26-8244-6181b21dece0\") " pod="openstack/horizon-648fb89557-mvw59" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.461091 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e3693fc-394d-4cbb-8501-34bbc0483c37-logs\") pod \"watcher-decision-engine-0\" (UID: \"8e3693fc-394d-4cbb-8501-34bbc0483c37\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.461124 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/8e3693fc-394d-4cbb-8501-34bbc0483c37-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"8e3693fc-394d-4cbb-8501-34bbc0483c37\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.461165 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqkgk\" (UniqueName: \"kubernetes.io/projected/8e3693fc-394d-4cbb-8501-34bbc0483c37-kube-api-access-xqkgk\") pod \"watcher-decision-engine-0\" (UID: \"8e3693fc-394d-4cbb-8501-34bbc0483c37\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.461200 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e3693fc-394d-4cbb-8501-34bbc0483c37-config-data\") pod \"watcher-decision-engine-0\" (UID: \"8e3693fc-394d-4cbb-8501-34bbc0483c37\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.461242 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2x5cz\" (UniqueName: \"kubernetes.io/projected/53f87c55-ecee-4d26-8244-6181b21dece0-kube-api-access-2x5cz\") pod \"horizon-648fb89557-mvw59\" (UID: \"53f87c55-ecee-4d26-8244-6181b21dece0\") " pod="openstack/horizon-648fb89557-mvw59" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.461261 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53f87c55-ecee-4d26-8244-6181b21dece0-scripts\") pod \"horizon-648fb89557-mvw59\" (UID: \"53f87c55-ecee-4d26-8244-6181b21dece0\") " pod="openstack/horizon-648fb89557-mvw59" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.461316 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e3693fc-394d-4cbb-8501-34bbc0483c37-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"8e3693fc-394d-4cbb-8501-34bbc0483c37\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.461333 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53f87c55-ecee-4d26-8244-6181b21dece0-logs\") pod \"horizon-648fb89557-mvw59\" (UID: \"53f87c55-ecee-4d26-8244-6181b21dece0\") " pod="openstack/horizon-648fb89557-mvw59" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.461373 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/53f87c55-ecee-4d26-8244-6181b21dece0-horizon-secret-key\") pod \"horizon-648fb89557-mvw59\" (UID: \"53f87c55-ecee-4d26-8244-6181b21dece0\") " pod="openstack/horizon-648fb89557-mvw59" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.462179 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53f87c55-ecee-4d26-8244-6181b21dece0-scripts\") pod \"horizon-648fb89557-mvw59\" (UID: \"53f87c55-ecee-4d26-8244-6181b21dece0\") " pod="openstack/horizon-648fb89557-mvw59" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.462231 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53f87c55-ecee-4d26-8244-6181b21dece0-config-data\") pod \"horizon-648fb89557-mvw59\" (UID: \"53f87c55-ecee-4d26-8244-6181b21dece0\") " pod="openstack/horizon-648fb89557-mvw59" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.462595 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53f87c55-ecee-4d26-8244-6181b21dece0-logs\") pod \"horizon-648fb89557-mvw59\" (UID: \"53f87c55-ecee-4d26-8244-6181b21dece0\") " pod="openstack/horizon-648fb89557-mvw59" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.465473 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2g9pt"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.473505 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/53f87c55-ecee-4d26-8244-6181b21dece0-horizon-secret-key\") pod \"horizon-648fb89557-mvw59\" (UID: \"53f87c55-ecee-4d26-8244-6181b21dece0\") " pod="openstack/horizon-648fb89557-mvw59" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.486363 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2x5cz\" (UniqueName: \"kubernetes.io/projected/53f87c55-ecee-4d26-8244-6181b21dece0-kube-api-access-2x5cz\") pod \"horizon-648fb89557-mvw59\" (UID: \"53f87c55-ecee-4d26-8244-6181b21dece0\") " pod="openstack/horizon-648fb89557-mvw59" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.509019 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8lz9j" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.516898 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-w8k5f"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.532523 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-w8k5f" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.535185 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-clw2v" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.538588 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.553702 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-w8k5f"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.555911 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-648fb89557-mvw59" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.562931 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/8e3693fc-394d-4cbb-8501-34bbc0483c37-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"8e3693fc-394d-4cbb-8501-34bbc0483c37\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.562992 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggth6\" (UniqueName: \"kubernetes.io/projected/ac00cda5-18ab-48de-a82e-1ebe08fa518c-kube-api-access-ggth6\") pod \"neutron-db-sync-2g9pt\" (UID: \"ac00cda5-18ab-48de-a82e-1ebe08fa518c\") " pod="openstack/neutron-db-sync-2g9pt" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.563012 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqkgk\" (UniqueName: \"kubernetes.io/projected/8e3693fc-394d-4cbb-8501-34bbc0483c37-kube-api-access-xqkgk\") pod \"watcher-decision-engine-0\" (UID: \"8e3693fc-394d-4cbb-8501-34bbc0483c37\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.563047 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e3693fc-394d-4cbb-8501-34bbc0483c37-config-data\") pod \"watcher-decision-engine-0\" (UID: \"8e3693fc-394d-4cbb-8501-34bbc0483c37\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.563085 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ac00cda5-18ab-48de-a82e-1ebe08fa518c-config\") pod \"neutron-db-sync-2g9pt\" (UID: \"ac00cda5-18ab-48de-a82e-1ebe08fa518c\") " pod="openstack/neutron-db-sync-2g9pt" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.563120 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e3693fc-394d-4cbb-8501-34bbc0483c37-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"8e3693fc-394d-4cbb-8501-34bbc0483c37\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.563161 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac00cda5-18ab-48de-a82e-1ebe08fa518c-combined-ca-bundle\") pod \"neutron-db-sync-2g9pt\" (UID: \"ac00cda5-18ab-48de-a82e-1ebe08fa518c\") " pod="openstack/neutron-db-sync-2g9pt" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.563185 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e3693fc-394d-4cbb-8501-34bbc0483c37-logs\") pod \"watcher-decision-engine-0\" (UID: \"8e3693fc-394d-4cbb-8501-34bbc0483c37\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.563814 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e3693fc-394d-4cbb-8501-34bbc0483c37-logs\") pod \"watcher-decision-engine-0\" (UID: \"8e3693fc-394d-4cbb-8501-34bbc0483c37\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.568479 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/8e3693fc-394d-4cbb-8501-34bbc0483c37-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"8e3693fc-394d-4cbb-8501-34bbc0483c37\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.573317 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e3693fc-394d-4cbb-8501-34bbc0483c37-config-data\") pod \"watcher-decision-engine-0\" (UID: \"8e3693fc-394d-4cbb-8501-34bbc0483c37\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.584371 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e3693fc-394d-4cbb-8501-34bbc0483c37-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"8e3693fc-394d-4cbb-8501-34bbc0483c37\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.591905 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ff6fd6c5-fnwm8"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.596617 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hd2nb" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.610494 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-67bd586fc-x2g2p"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.611862 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67bd586fc-x2g2p" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.622641 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqkgk\" (UniqueName: \"kubernetes.io/projected/8e3693fc-394d-4cbb-8501-34bbc0483c37-kube-api-access-xqkgk\") pod \"watcher-decision-engine-0\" (UID: \"8e3693fc-394d-4cbb-8501-34bbc0483c37\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.628257 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-dd959b98c-6ffpz"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.630357 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.669885 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac00cda5-18ab-48de-a82e-1ebe08fa518c-combined-ca-bundle\") pod \"neutron-db-sync-2g9pt\" (UID: \"ac00cda5-18ab-48de-a82e-1ebe08fa518c\") " pod="openstack/neutron-db-sync-2g9pt" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.669977 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfq6t\" (UniqueName: \"kubernetes.io/projected/5d2429d8-9baa-4532-9f8b-9ffb72a3a764-kube-api-access-pfq6t\") pod \"barbican-db-sync-w8k5f\" (UID: \"5d2429d8-9baa-4532-9f8b-9ffb72a3a764\") " pod="openstack/barbican-db-sync-w8k5f" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.670013 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5d2429d8-9baa-4532-9f8b-9ffb72a3a764-db-sync-config-data\") pod \"barbican-db-sync-w8k5f\" (UID: \"5d2429d8-9baa-4532-9f8b-9ffb72a3a764\") " pod="openstack/barbican-db-sync-w8k5f" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.670038 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggth6\" (UniqueName: \"kubernetes.io/projected/ac00cda5-18ab-48de-a82e-1ebe08fa518c-kube-api-access-ggth6\") pod \"neutron-db-sync-2g9pt\" (UID: \"ac00cda5-18ab-48de-a82e-1ebe08fa518c\") " pod="openstack/neutron-db-sync-2g9pt" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.670750 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ac00cda5-18ab-48de-a82e-1ebe08fa518c-config\") pod \"neutron-db-sync-2g9pt\" (UID: \"ac00cda5-18ab-48de-a82e-1ebe08fa518c\") " pod="openstack/neutron-db-sync-2g9pt" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.670798 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d2429d8-9baa-4532-9f8b-9ffb72a3a764-combined-ca-bundle\") pod \"barbican-db-sync-w8k5f\" (UID: \"5d2429d8-9baa-4532-9f8b-9ffb72a3a764\") " pod="openstack/barbican-db-sync-w8k5f" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.720383 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggth6\" (UniqueName: \"kubernetes.io/projected/ac00cda5-18ab-48de-a82e-1ebe08fa518c-kube-api-access-ggth6\") pod \"neutron-db-sync-2g9pt\" (UID: \"ac00cda5-18ab-48de-a82e-1ebe08fa518c\") " pod="openstack/neutron-db-sync-2g9pt" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.740931 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ac00cda5-18ab-48de-a82e-1ebe08fa518c-config\") pod \"neutron-db-sync-2g9pt\" (UID: \"ac00cda5-18ab-48de-a82e-1ebe08fa518c\") " pod="openstack/neutron-db-sync-2g9pt" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.741478 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac00cda5-18ab-48de-a82e-1ebe08fa518c-combined-ca-bundle\") pod \"neutron-db-sync-2g9pt\" (UID: \"ac00cda5-18ab-48de-a82e-1ebe08fa518c\") " pod="openstack/neutron-db-sync-2g9pt" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.779418 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.780461 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gcsz\" (UniqueName: \"kubernetes.io/projected/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-kube-api-access-9gcsz\") pod \"dnsmasq-dns-dd959b98c-6ffpz\" (UID: \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\") " pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.780514 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d2429d8-9baa-4532-9f8b-9ffb72a3a764-combined-ca-bundle\") pod \"barbican-db-sync-w8k5f\" (UID: \"5d2429d8-9baa-4532-9f8b-9ffb72a3a764\") " pod="openstack/barbican-db-sync-w8k5f" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.780536 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-ovsdbserver-sb\") pod \"dnsmasq-dns-dd959b98c-6ffpz\" (UID: \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\") " pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.780558 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef216407-b6ec-4691-8f49-cf8fce3f3544-scripts\") pod \"horizon-67bd586fc-x2g2p\" (UID: \"ef216407-b6ec-4691-8f49-cf8fce3f3544\") " pod="openstack/horizon-67bd586fc-x2g2p" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.780605 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-dns-swift-storage-0\") pod \"dnsmasq-dns-dd959b98c-6ffpz\" (UID: \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\") " pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.780626 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef216407-b6ec-4691-8f49-cf8fce3f3544-config-data\") pod \"horizon-67bd586fc-x2g2p\" (UID: \"ef216407-b6ec-4691-8f49-cf8fce3f3544\") " pod="openstack/horizon-67bd586fc-x2g2p" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.780648 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef216407-b6ec-4691-8f49-cf8fce3f3544-logs\") pod \"horizon-67bd586fc-x2g2p\" (UID: \"ef216407-b6ec-4691-8f49-cf8fce3f3544\") " pod="openstack/horizon-67bd586fc-x2g2p" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.780663 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlq82\" (UniqueName: \"kubernetes.io/projected/ef216407-b6ec-4691-8f49-cf8fce3f3544-kube-api-access-rlq82\") pod \"horizon-67bd586fc-x2g2p\" (UID: \"ef216407-b6ec-4691-8f49-cf8fce3f3544\") " pod="openstack/horizon-67bd586fc-x2g2p" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.780682 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-ovsdbserver-nb\") pod \"dnsmasq-dns-dd959b98c-6ffpz\" (UID: \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\") " pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.780700 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-dns-svc\") pod \"dnsmasq-dns-dd959b98c-6ffpz\" (UID: \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\") " pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.780722 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-config\") pod \"dnsmasq-dns-dd959b98c-6ffpz\" (UID: \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\") " pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.780742 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfq6t\" (UniqueName: \"kubernetes.io/projected/5d2429d8-9baa-4532-9f8b-9ffb72a3a764-kube-api-access-pfq6t\") pod \"barbican-db-sync-w8k5f\" (UID: \"5d2429d8-9baa-4532-9f8b-9ffb72a3a764\") " pod="openstack/barbican-db-sync-w8k5f" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.780780 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5d2429d8-9baa-4532-9f8b-9ffb72a3a764-db-sync-config-data\") pod \"barbican-db-sync-w8k5f\" (UID: \"5d2429d8-9baa-4532-9f8b-9ffb72a3a764\") " pod="openstack/barbican-db-sync-w8k5f" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.780806 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ef216407-b6ec-4691-8f49-cf8fce3f3544-horizon-secret-key\") pod \"horizon-67bd586fc-x2g2p\" (UID: \"ef216407-b6ec-4691-8f49-cf8fce3f3544\") " pod="openstack/horizon-67bd586fc-x2g2p" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.820327 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d2429d8-9baa-4532-9f8b-9ffb72a3a764-combined-ca-bundle\") pod \"barbican-db-sync-w8k5f\" (UID: \"5d2429d8-9baa-4532-9f8b-9ffb72a3a764\") " pod="openstack/barbican-db-sync-w8k5f" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.820545 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5d2429d8-9baa-4532-9f8b-9ffb72a3a764-db-sync-config-data\") pod \"barbican-db-sync-w8k5f\" (UID: \"5d2429d8-9baa-4532-9f8b-9ffb72a3a764\") " pod="openstack/barbican-db-sync-w8k5f" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.836407 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfq6t\" (UniqueName: \"kubernetes.io/projected/5d2429d8-9baa-4532-9f8b-9ffb72a3a764-kube-api-access-pfq6t\") pod \"barbican-db-sync-w8k5f\" (UID: \"5d2429d8-9baa-4532-9f8b-9ffb72a3a764\") " pod="openstack/barbican-db-sync-w8k5f" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.876502 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-67bd586fc-x2g2p"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.876571 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dd959b98c-6ffpz"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.876586 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-qw92l"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.878271 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.880173 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qw92l" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.880893 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-qw92l"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.880917 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.880994 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.881953 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-dns-swift-storage-0\") pod \"dnsmasq-dns-dd959b98c-6ffpz\" (UID: \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\") " pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.882055 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef216407-b6ec-4691-8f49-cf8fce3f3544-config-data\") pod \"horizon-67bd586fc-x2g2p\" (UID: \"ef216407-b6ec-4691-8f49-cf8fce3f3544\") " pod="openstack/horizon-67bd586fc-x2g2p" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.882126 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef216407-b6ec-4691-8f49-cf8fce3f3544-logs\") pod \"horizon-67bd586fc-x2g2p\" (UID: \"ef216407-b6ec-4691-8f49-cf8fce3f3544\") " pod="openstack/horizon-67bd586fc-x2g2p" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.882182 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlq82\" (UniqueName: \"kubernetes.io/projected/ef216407-b6ec-4691-8f49-cf8fce3f3544-kube-api-access-rlq82\") pod \"horizon-67bd586fc-x2g2p\" (UID: \"ef216407-b6ec-4691-8f49-cf8fce3f3544\") " pod="openstack/horizon-67bd586fc-x2g2p" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.882245 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-ovsdbserver-nb\") pod \"dnsmasq-dns-dd959b98c-6ffpz\" (UID: \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\") " pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.882316 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-dns-svc\") pod \"dnsmasq-dns-dd959b98c-6ffpz\" (UID: \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\") " pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.882401 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-config\") pod \"dnsmasq-dns-dd959b98c-6ffpz\" (UID: \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\") " pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.882488 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ef216407-b6ec-4691-8f49-cf8fce3f3544-horizon-secret-key\") pod \"horizon-67bd586fc-x2g2p\" (UID: \"ef216407-b6ec-4691-8f49-cf8fce3f3544\") " pod="openstack/horizon-67bd586fc-x2g2p" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.882559 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gcsz\" (UniqueName: \"kubernetes.io/projected/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-kube-api-access-9gcsz\") pod \"dnsmasq-dns-dd959b98c-6ffpz\" (UID: \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\") " pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.882648 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-ovsdbserver-sb\") pod \"dnsmasq-dns-dd959b98c-6ffpz\" (UID: \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\") " pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.882718 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef216407-b6ec-4691-8f49-cf8fce3f3544-scripts\") pod \"horizon-67bd586fc-x2g2p\" (UID: \"ef216407-b6ec-4691-8f49-cf8fce3f3544\") " pod="openstack/horizon-67bd586fc-x2g2p" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.883179 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.883430 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-9r2gs" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.883537 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.883888 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef216407-b6ec-4691-8f49-cf8fce3f3544-scripts\") pod \"horizon-67bd586fc-x2g2p\" (UID: \"ef216407-b6ec-4691-8f49-cf8fce3f3544\") " pod="openstack/horizon-67bd586fc-x2g2p" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.884453 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-ovsdbserver-nb\") pod \"dnsmasq-dns-dd959b98c-6ffpz\" (UID: \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\") " pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.884540 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-dns-svc\") pod \"dnsmasq-dns-dd959b98c-6ffpz\" (UID: \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\") " pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.885654 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-dns-swift-storage-0\") pod \"dnsmasq-dns-dd959b98c-6ffpz\" (UID: \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\") " pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.886023 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef216407-b6ec-4691-8f49-cf8fce3f3544-logs\") pod \"horizon-67bd586fc-x2g2p\" (UID: \"ef216407-b6ec-4691-8f49-cf8fce3f3544\") " pod="openstack/horizon-67bd586fc-x2g2p" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.886621 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef216407-b6ec-4691-8f49-cf8fce3f3544-config-data\") pod \"horizon-67bd586fc-x2g2p\" (UID: \"ef216407-b6ec-4691-8f49-cf8fce3f3544\") " pod="openstack/horizon-67bd586fc-x2g2p" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.886650 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-config\") pod \"dnsmasq-dns-dd959b98c-6ffpz\" (UID: \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\") " pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.889677 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.891088 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ef216407-b6ec-4691-8f49-cf8fce3f3544-horizon-secret-key\") pod \"horizon-67bd586fc-x2g2p\" (UID: \"ef216407-b6ec-4691-8f49-cf8fce3f3544\") " pod="openstack/horizon-67bd586fc-x2g2p" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.890720 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.892714 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-ovsdbserver-sb\") pod \"dnsmasq-dns-dd959b98c-6ffpz\" (UID: \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\") " pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.916780 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gcsz\" (UniqueName: \"kubernetes.io/projected/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-kube-api-access-9gcsz\") pod \"dnsmasq-dns-dd959b98c-6ffpz\" (UID: \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\") " pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.920536 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlq82\" (UniqueName: \"kubernetes.io/projected/ef216407-b6ec-4691-8f49-cf8fce3f3544-kube-api-access-rlq82\") pod \"horizon-67bd586fc-x2g2p\" (UID: \"ef216407-b6ec-4691-8f49-cf8fce3f3544\") " pod="openstack/horizon-67bd586fc-x2g2p" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.971977 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2g9pt" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.987423 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cca43ea-2244-4f5c-8704-8d91634c0ad6-config-data\") pod \"placement-db-sync-qw92l\" (UID: \"5cca43ea-2244-4f5c-8704-8d91634c0ad6\") " pod="openstack/placement-db-sync-qw92l" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.987681 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cca43ea-2244-4f5c-8704-8d91634c0ad6-combined-ca-bundle\") pod \"placement-db-sync-qw92l\" (UID: \"5cca43ea-2244-4f5c-8704-8d91634c0ad6\") " pod="openstack/placement-db-sync-qw92l" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.987729 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e4faa24a-a43f-4d24-b20d-09e241d0cc96-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " pod="openstack/ceilometer-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.987819 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4faa24a-a43f-4d24-b20d-09e241d0cc96-scripts\") pod \"ceilometer-0\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " pod="openstack/ceilometer-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.987838 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4faa24a-a43f-4d24-b20d-09e241d0cc96-config-data\") pod \"ceilometer-0\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " pod="openstack/ceilometer-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.987920 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6j4p\" (UniqueName: \"kubernetes.io/projected/e4faa24a-a43f-4d24-b20d-09e241d0cc96-kube-api-access-n6j4p\") pod \"ceilometer-0\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " pod="openstack/ceilometer-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.987941 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4faa24a-a43f-4d24-b20d-09e241d0cc96-log-httpd\") pod \"ceilometer-0\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " pod="openstack/ceilometer-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.987971 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cca43ea-2244-4f5c-8704-8d91634c0ad6-logs\") pod \"placement-db-sync-qw92l\" (UID: \"5cca43ea-2244-4f5c-8704-8d91634c0ad6\") " pod="openstack/placement-db-sync-qw92l" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.987998 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cca43ea-2244-4f5c-8704-8d91634c0ad6-scripts\") pod \"placement-db-sync-qw92l\" (UID: \"5cca43ea-2244-4f5c-8704-8d91634c0ad6\") " pod="openstack/placement-db-sync-qw92l" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.988048 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk4d5\" (UniqueName: \"kubernetes.io/projected/5cca43ea-2244-4f5c-8704-8d91634c0ad6-kube-api-access-mk4d5\") pod \"placement-db-sync-qw92l\" (UID: \"5cca43ea-2244-4f5c-8704-8d91634c0ad6\") " pod="openstack/placement-db-sync-qw92l" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.988095 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4faa24a-a43f-4d24-b20d-09e241d0cc96-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " pod="openstack/ceilometer-0" Dec 03 16:23:57 crc kubenswrapper[4998]: I1203 16:23:57.988196 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4faa24a-a43f-4d24-b20d-09e241d0cc96-run-httpd\") pod \"ceilometer-0\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " pod="openstack/ceilometer-0" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:57.999724 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-w8k5f" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.016258 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67bd586fc-x2g2p" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.041461 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.089748 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cca43ea-2244-4f5c-8704-8d91634c0ad6-combined-ca-bundle\") pod \"placement-db-sync-qw92l\" (UID: \"5cca43ea-2244-4f5c-8704-8d91634c0ad6\") " pod="openstack/placement-db-sync-qw92l" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.089915 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e4faa24a-a43f-4d24-b20d-09e241d0cc96-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " pod="openstack/ceilometer-0" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.089989 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4faa24a-a43f-4d24-b20d-09e241d0cc96-scripts\") pod \"ceilometer-0\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " pod="openstack/ceilometer-0" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.090006 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4faa24a-a43f-4d24-b20d-09e241d0cc96-config-data\") pod \"ceilometer-0\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " pod="openstack/ceilometer-0" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.090319 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6j4p\" (UniqueName: \"kubernetes.io/projected/e4faa24a-a43f-4d24-b20d-09e241d0cc96-kube-api-access-n6j4p\") pod \"ceilometer-0\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " pod="openstack/ceilometer-0" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.090352 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4faa24a-a43f-4d24-b20d-09e241d0cc96-log-httpd\") pod \"ceilometer-0\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " pod="openstack/ceilometer-0" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.090376 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cca43ea-2244-4f5c-8704-8d91634c0ad6-logs\") pod \"placement-db-sync-qw92l\" (UID: \"5cca43ea-2244-4f5c-8704-8d91634c0ad6\") " pod="openstack/placement-db-sync-qw92l" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.090434 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cca43ea-2244-4f5c-8704-8d91634c0ad6-scripts\") pod \"placement-db-sync-qw92l\" (UID: \"5cca43ea-2244-4f5c-8704-8d91634c0ad6\") " pod="openstack/placement-db-sync-qw92l" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.090460 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk4d5\" (UniqueName: \"kubernetes.io/projected/5cca43ea-2244-4f5c-8704-8d91634c0ad6-kube-api-access-mk4d5\") pod \"placement-db-sync-qw92l\" (UID: \"5cca43ea-2244-4f5c-8704-8d91634c0ad6\") " pod="openstack/placement-db-sync-qw92l" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.090507 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4faa24a-a43f-4d24-b20d-09e241d0cc96-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " pod="openstack/ceilometer-0" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.090526 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4faa24a-a43f-4d24-b20d-09e241d0cc96-run-httpd\") pod \"ceilometer-0\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " pod="openstack/ceilometer-0" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.091173 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cca43ea-2244-4f5c-8704-8d91634c0ad6-config-data\") pod \"placement-db-sync-qw92l\" (UID: \"5cca43ea-2244-4f5c-8704-8d91634c0ad6\") " pod="openstack/placement-db-sync-qw92l" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.092403 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4faa24a-a43f-4d24-b20d-09e241d0cc96-log-httpd\") pod \"ceilometer-0\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " pod="openstack/ceilometer-0" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.093151 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cca43ea-2244-4f5c-8704-8d91634c0ad6-logs\") pod \"placement-db-sync-qw92l\" (UID: \"5cca43ea-2244-4f5c-8704-8d91634c0ad6\") " pod="openstack/placement-db-sync-qw92l" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.093454 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4faa24a-a43f-4d24-b20d-09e241d0cc96-run-httpd\") pod \"ceilometer-0\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " pod="openstack/ceilometer-0" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.095205 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4faa24a-a43f-4d24-b20d-09e241d0cc96-scripts\") pod \"ceilometer-0\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " pod="openstack/ceilometer-0" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.096334 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4faa24a-a43f-4d24-b20d-09e241d0cc96-config-data\") pod \"ceilometer-0\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " pod="openstack/ceilometer-0" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.097124 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e4faa24a-a43f-4d24-b20d-09e241d0cc96-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " pod="openstack/ceilometer-0" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.097739 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cca43ea-2244-4f5c-8704-8d91634c0ad6-config-data\") pod \"placement-db-sync-qw92l\" (UID: \"5cca43ea-2244-4f5c-8704-8d91634c0ad6\") " pod="openstack/placement-db-sync-qw92l" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.098463 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4faa24a-a43f-4d24-b20d-09e241d0cc96-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " pod="openstack/ceilometer-0" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.098691 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cca43ea-2244-4f5c-8704-8d91634c0ad6-scripts\") pod \"placement-db-sync-qw92l\" (UID: \"5cca43ea-2244-4f5c-8704-8d91634c0ad6\") " pod="openstack/placement-db-sync-qw92l" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.099025 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cca43ea-2244-4f5c-8704-8d91634c0ad6-combined-ca-bundle\") pod \"placement-db-sync-qw92l\" (UID: \"5cca43ea-2244-4f5c-8704-8d91634c0ad6\") " pod="openstack/placement-db-sync-qw92l" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.113200 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6j4p\" (UniqueName: \"kubernetes.io/projected/e4faa24a-a43f-4d24-b20d-09e241d0cc96-kube-api-access-n6j4p\") pod \"ceilometer-0\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " pod="openstack/ceilometer-0" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.115802 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk4d5\" (UniqueName: \"kubernetes.io/projected/5cca43ea-2244-4f5c-8704-8d91634c0ad6-kube-api-access-mk4d5\") pod \"placement-db-sync-qw92l\" (UID: \"5cca43ea-2244-4f5c-8704-8d91634c0ad6\") " pod="openstack/placement-db-sync-qw92l" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.357962 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.358459 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ff6fd6c5-fnwm8"] Dec 03 16:23:58 crc kubenswrapper[4998]: W1203 16:23:58.365203 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd1aa1ba_fae8_4b45_8af0_a83948b2dd12.slice/crio-2eb0f689f0284e70a7b6250a4451664f64e3dbeadd783ad55b9a090237382dac WatchSource:0}: Error finding container 2eb0f689f0284e70a7b6250a4451664f64e3dbeadd783ad55b9a090237382dac: Status 404 returned error can't find the container with id 2eb0f689f0284e70a7b6250a4451664f64e3dbeadd783ad55b9a090237382dac Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.371978 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.377676 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qw92l" Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.472700 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.839250 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-8lz9j"] Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.870976 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.897144 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-648fb89557-mvw59"] Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.945714 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"e968cb62-c964-46f9-ac6d-6cb674c8bf22","Type":"ContainerStarted","Data":"3b071df7491234fd479ab5af697909c349b263dc051375c3ab6f1c89a97bad20"} Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.945772 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"e968cb62-c964-46f9-ac6d-6cb674c8bf22","Type":"ContainerStarted","Data":"fb0d0504f1fffae1f15467db418cfa442ac6dfcd08ea82fc04ccd61368a5ae9c"} Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.952763 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"5b94f565-0948-4e4f-9636-2833d43bb829","Type":"ContainerStarted","Data":"70609850792f79e03427fa3205636c4090b8ec99876d05763ae501a7d07fb7a6"} Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.957692 4998 generic.go:334] "Generic (PLEG): container finished" podID="dd1aa1ba-fae8-4b45-8af0-a83948b2dd12" containerID="f415919649a9e111efcc95b1b133997e7f4e5a9126d4af55f2518a8d14a4619f" exitCode=0 Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.957729 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ff6fd6c5-fnwm8" event={"ID":"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12","Type":"ContainerDied","Data":"f415919649a9e111efcc95b1b133997e7f4e5a9126d4af55f2518a8d14a4619f"} Dec 03 16:23:58 crc kubenswrapper[4998]: I1203 16:23:58.957765 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ff6fd6c5-fnwm8" event={"ID":"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12","Type":"ContainerStarted","Data":"2eb0f689f0284e70a7b6250a4451664f64e3dbeadd783ad55b9a090237382dac"} Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.012874 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-w8k5f"] Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.038076 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2g9pt"] Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.044949 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-67bd586fc-x2g2p"] Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.078855 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dd959b98c-6ffpz"] Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.094065 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hd2nb"] Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.203248 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-qw92l"] Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.268237 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.361413 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.380192 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-648fb89557-mvw59"] Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.477906 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7f44dc4fd9-2rfng"] Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.480809 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f44dc4fd9-2rfng" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.488994 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f44dc4fd9-2rfng"] Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.595874 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.656577 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ff6fd6c5-fnwm8" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.667510 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwlcw\" (UniqueName: \"kubernetes.io/projected/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-kube-api-access-lwlcw\") pod \"horizon-7f44dc4fd9-2rfng\" (UID: \"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f\") " pod="openstack/horizon-7f44dc4fd9-2rfng" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.667706 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-horizon-secret-key\") pod \"horizon-7f44dc4fd9-2rfng\" (UID: \"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f\") " pod="openstack/horizon-7f44dc4fd9-2rfng" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.667986 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-scripts\") pod \"horizon-7f44dc4fd9-2rfng\" (UID: \"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f\") " pod="openstack/horizon-7f44dc4fd9-2rfng" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.668033 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-logs\") pod \"horizon-7f44dc4fd9-2rfng\" (UID: \"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f\") " pod="openstack/horizon-7f44dc4fd9-2rfng" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.668207 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-config-data\") pod \"horizon-7f44dc4fd9-2rfng\" (UID: \"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f\") " pod="openstack/horizon-7f44dc4fd9-2rfng" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.771806 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-ovsdbserver-nb\") pod \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\" (UID: \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\") " Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.771845 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-config\") pod \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\" (UID: \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\") " Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.771916 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-ovsdbserver-sb\") pod \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\" (UID: \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\") " Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.772050 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dj2x6\" (UniqueName: \"kubernetes.io/projected/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-kube-api-access-dj2x6\") pod \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\" (UID: \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\") " Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.772111 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-dns-svc\") pod \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\" (UID: \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\") " Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.772151 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-dns-swift-storage-0\") pod \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\" (UID: \"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12\") " Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.772462 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-horizon-secret-key\") pod \"horizon-7f44dc4fd9-2rfng\" (UID: \"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f\") " pod="openstack/horizon-7f44dc4fd9-2rfng" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.772507 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-scripts\") pod \"horizon-7f44dc4fd9-2rfng\" (UID: \"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f\") " pod="openstack/horizon-7f44dc4fd9-2rfng" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.772527 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-logs\") pod \"horizon-7f44dc4fd9-2rfng\" (UID: \"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f\") " pod="openstack/horizon-7f44dc4fd9-2rfng" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.772546 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-config-data\") pod \"horizon-7f44dc4fd9-2rfng\" (UID: \"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f\") " pod="openstack/horizon-7f44dc4fd9-2rfng" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.772587 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwlcw\" (UniqueName: \"kubernetes.io/projected/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-kube-api-access-lwlcw\") pod \"horizon-7f44dc4fd9-2rfng\" (UID: \"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f\") " pod="openstack/horizon-7f44dc4fd9-2rfng" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.777928 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-scripts\") pod \"horizon-7f44dc4fd9-2rfng\" (UID: \"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f\") " pod="openstack/horizon-7f44dc4fd9-2rfng" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.778174 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-logs\") pod \"horizon-7f44dc4fd9-2rfng\" (UID: \"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f\") " pod="openstack/horizon-7f44dc4fd9-2rfng" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.779967 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-config-data\") pod \"horizon-7f44dc4fd9-2rfng\" (UID: \"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f\") " pod="openstack/horizon-7f44dc4fd9-2rfng" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.782394 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-kube-api-access-dj2x6" (OuterVolumeSpecName: "kube-api-access-dj2x6") pod "dd1aa1ba-fae8-4b45-8af0-a83948b2dd12" (UID: "dd1aa1ba-fae8-4b45-8af0-a83948b2dd12"). InnerVolumeSpecName "kube-api-access-dj2x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.799124 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-horizon-secret-key\") pod \"horizon-7f44dc4fd9-2rfng\" (UID: \"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f\") " pod="openstack/horizon-7f44dc4fd9-2rfng" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.819102 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwlcw\" (UniqueName: \"kubernetes.io/projected/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-kube-api-access-lwlcw\") pod \"horizon-7f44dc4fd9-2rfng\" (UID: \"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f\") " pod="openstack/horizon-7f44dc4fd9-2rfng" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.843250 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-config" (OuterVolumeSpecName: "config") pod "dd1aa1ba-fae8-4b45-8af0-a83948b2dd12" (UID: "dd1aa1ba-fae8-4b45-8af0-a83948b2dd12"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.843613 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "dd1aa1ba-fae8-4b45-8af0-a83948b2dd12" (UID: "dd1aa1ba-fae8-4b45-8af0-a83948b2dd12"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.851739 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "dd1aa1ba-fae8-4b45-8af0-a83948b2dd12" (UID: "dd1aa1ba-fae8-4b45-8af0-a83948b2dd12"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.853301 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dd1aa1ba-fae8-4b45-8af0-a83948b2dd12" (UID: "dd1aa1ba-fae8-4b45-8af0-a83948b2dd12"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.875848 4998 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.875888 4998 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.875899 4998 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.875912 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.875931 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dj2x6\" (UniqueName: \"kubernetes.io/projected/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-kube-api-access-dj2x6\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.880255 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "dd1aa1ba-fae8-4b45-8af0-a83948b2dd12" (UID: "dd1aa1ba-fae8-4b45-8af0-a83948b2dd12"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.946781 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f44dc4fd9-2rfng" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.978002 4998 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 16:23:59 crc kubenswrapper[4998]: I1203 16:23:59.998090 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-648fb89557-mvw59" event={"ID":"53f87c55-ecee-4d26-8244-6181b21dece0","Type":"ContainerStarted","Data":"07e18fb60fe608ea212b8c98f530c643287fbdff81593c6a34fefa9c5b99e086"} Dec 03 16:24:00 crc kubenswrapper[4998]: I1203 16:24:00.005210 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8lz9j" event={"ID":"82aa9d23-457c-45a0-9729-e5a803738237","Type":"ContainerStarted","Data":"f5c3e814f3475f33ec9ee3b208a68df32ef8b689c79d08eff56c919ab7090191"} Dec 03 16:24:00 crc kubenswrapper[4998]: I1203 16:24:00.036555 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ff6fd6c5-fnwm8" Dec 03 16:24:00 crc kubenswrapper[4998]: I1203 16:24:00.037388 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ff6fd6c5-fnwm8" event={"ID":"dd1aa1ba-fae8-4b45-8af0-a83948b2dd12","Type":"ContainerDied","Data":"2eb0f689f0284e70a7b6250a4451664f64e3dbeadd783ad55b9a090237382dac"} Dec 03 16:24:00 crc kubenswrapper[4998]: I1203 16:24:00.037438 4998 scope.go:117] "RemoveContainer" containerID="f415919649a9e111efcc95b1b133997e7f4e5a9126d4af55f2518a8d14a4619f" Dec 03 16:24:00 crc kubenswrapper[4998]: I1203 16:24:00.057787 4998 generic.go:334] "Generic (PLEG): container finished" podID="6f1c1a9e-4b7e-474a-99c9-0b79b259f774" containerID="e10ef0292f2e0c2ec6385af6e9ca41e1eb4a1620a7c75659d92488e45404d91c" exitCode=0 Dec 03 16:24:00 crc kubenswrapper[4998]: I1203 16:24:00.057841 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" event={"ID":"6f1c1a9e-4b7e-474a-99c9-0b79b259f774","Type":"ContainerDied","Data":"e10ef0292f2e0c2ec6385af6e9ca41e1eb4a1620a7c75659d92488e45404d91c"} Dec 03 16:24:00 crc kubenswrapper[4998]: I1203 16:24:00.057862 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" event={"ID":"6f1c1a9e-4b7e-474a-99c9-0b79b259f774","Type":"ContainerStarted","Data":"986b94e956c674309235ca750ba3b6e7573e938fe15d44013cb5fb07b9c49c76"} Dec 03 16:24:00 crc kubenswrapper[4998]: I1203 16:24:00.072804 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-w8k5f" event={"ID":"5d2429d8-9baa-4532-9f8b-9ffb72a3a764","Type":"ContainerStarted","Data":"f7b8c0dcf9cc5974b58aa07b9c714976da85b6a9e3adee1828e512e498b2c95f"} Dec 03 16:24:00 crc kubenswrapper[4998]: I1203 16:24:00.075873 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"e968cb62-c964-46f9-ac6d-6cb674c8bf22","Type":"ContainerStarted","Data":"2e5f63217f2477382afa7cfbebe41754b712ad2a5e18028b07c3c24bd16131f4"} Dec 03 16:24:00 crc kubenswrapper[4998]: I1203 16:24:00.076322 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Dec 03 16:24:00 crc kubenswrapper[4998]: I1203 16:24:00.088180 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67bd586fc-x2g2p" event={"ID":"ef216407-b6ec-4691-8f49-cf8fce3f3544","Type":"ContainerStarted","Data":"683c721d22b4678bf597f95fb591e6a027d68e3052836353969ee8d7e7a863ba"} Dec 03 16:24:00 crc kubenswrapper[4998]: I1203 16:24:00.111047 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ff6fd6c5-fnwm8"] Dec 03 16:24:00 crc kubenswrapper[4998]: I1203 16:24:00.117559 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"8e3693fc-394d-4cbb-8501-34bbc0483c37","Type":"ContainerStarted","Data":"0953a810f801d2d84d0ac7482bf35710bf063e49da93bce33dc04703dab5e740"} Dec 03 16:24:00 crc kubenswrapper[4998]: I1203 16:24:00.118985 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ff6fd6c5-fnwm8"] Dec 03 16:24:00 crc kubenswrapper[4998]: I1203 16:24:00.171888 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=4.171867453 podStartE2EDuration="4.171867453s" podCreationTimestamp="2025-12-03 16:23:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:24:00.162953528 +0000 UTC m=+1218.774653771" watchObservedRunningTime="2025-12-03 16:24:00.171867453 +0000 UTC m=+1218.783567676" Dec 03 16:24:00 crc kubenswrapper[4998]: I1203 16:24:00.184832 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qw92l" event={"ID":"5cca43ea-2244-4f5c-8704-8d91634c0ad6","Type":"ContainerStarted","Data":"068164d44989316e3d60e0b6bc2fde5aaabb737063cf017e0f9900b141460798"} Dec 03 16:24:00 crc kubenswrapper[4998]: I1203 16:24:00.190435 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4faa24a-a43f-4d24-b20d-09e241d0cc96","Type":"ContainerStarted","Data":"0475afca48189c206af41d74b245bf5418653f1ca935fd76768cd39cef6f38b0"} Dec 03 16:24:00 crc kubenswrapper[4998]: I1203 16:24:00.193309 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hd2nb" event={"ID":"aa92f53b-42a5-4b6d-aba4-3bec5917cd59","Type":"ContainerStarted","Data":"dce18792baccb8a737a1d6afcfdeae6a5b72eb876a08b9697f50acde4e2584ba"} Dec 03 16:24:00 crc kubenswrapper[4998]: I1203 16:24:00.193348 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hd2nb" event={"ID":"aa92f53b-42a5-4b6d-aba4-3bec5917cd59","Type":"ContainerStarted","Data":"ba4c8297559225b3f4f9e9050c9473293a0139681e97a465e39cd122e444b222"} Dec 03 16:24:00 crc kubenswrapper[4998]: I1203 16:24:00.198565 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2g9pt" event={"ID":"ac00cda5-18ab-48de-a82e-1ebe08fa518c","Type":"ContainerStarted","Data":"d9de792d4b5c1b7f62defd70fa80c00e1a269c701e996345d575192a2ef656f2"} Dec 03 16:24:00 crc kubenswrapper[4998]: I1203 16:24:00.199931 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2g9pt" event={"ID":"ac00cda5-18ab-48de-a82e-1ebe08fa518c","Type":"ContainerStarted","Data":"24f750ec22c49315b4d52610f475929dea6e8d7ec23668f695b057b4182a54ef"} Dec 03 16:24:00 crc kubenswrapper[4998]: I1203 16:24:00.219592 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-hd2nb" podStartSLOduration=4.219573774 podStartE2EDuration="4.219573774s" podCreationTimestamp="2025-12-03 16:23:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:24:00.217499282 +0000 UTC m=+1218.829199505" watchObservedRunningTime="2025-12-03 16:24:00.219573774 +0000 UTC m=+1218.831273997" Dec 03 16:24:00 crc kubenswrapper[4998]: I1203 16:24:00.239176 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-2g9pt" podStartSLOduration=3.239158558 podStartE2EDuration="3.239158558s" podCreationTimestamp="2025-12-03 16:23:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:24:00.238213624 +0000 UTC m=+1218.849913857" watchObservedRunningTime="2025-12-03 16:24:00.239158558 +0000 UTC m=+1218.850858781" Dec 03 16:24:00 crc kubenswrapper[4998]: I1203 16:24:00.550575 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f44dc4fd9-2rfng"] Dec 03 16:24:01 crc kubenswrapper[4998]: I1203 16:24:01.217390 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" event={"ID":"6f1c1a9e-4b7e-474a-99c9-0b79b259f774","Type":"ContainerStarted","Data":"e4dd17f34f33175e96a5ce0731b4953eafffeff8eb564be760614bdde21c1e9c"} Dec 03 16:24:01 crc kubenswrapper[4998]: I1203 16:24:01.217809 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="e968cb62-c964-46f9-ac6d-6cb674c8bf22" containerName="watcher-api-log" containerID="cri-o://3b071df7491234fd479ab5af697909c349b263dc051375c3ab6f1c89a97bad20" gracePeriod=30 Dec 03 16:24:01 crc kubenswrapper[4998]: I1203 16:24:01.218148 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="e968cb62-c964-46f9-ac6d-6cb674c8bf22" containerName="watcher-api" containerID="cri-o://2e5f63217f2477382afa7cfbebe41754b712ad2a5e18028b07c3c24bd16131f4" gracePeriod=30 Dec 03 16:24:01 crc kubenswrapper[4998]: I1203 16:24:01.220257 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" Dec 03 16:24:01 crc kubenswrapper[4998]: I1203 16:24:01.228360 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="e968cb62-c964-46f9-ac6d-6cb674c8bf22" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.147:9322/\": EOF" Dec 03 16:24:01 crc kubenswrapper[4998]: I1203 16:24:01.249705 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" podStartSLOduration=4.249690116 podStartE2EDuration="4.249690116s" podCreationTimestamp="2025-12-03 16:23:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:24:01.248060556 +0000 UTC m=+1219.859760779" watchObservedRunningTime="2025-12-03 16:24:01.249690116 +0000 UTC m=+1219.861390339" Dec 03 16:24:01 crc kubenswrapper[4998]: I1203 16:24:01.700347 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd1aa1ba-fae8-4b45-8af0-a83948b2dd12" path="/var/lib/kubelet/pods/dd1aa1ba-fae8-4b45-8af0-a83948b2dd12/volumes" Dec 03 16:24:02 crc kubenswrapper[4998]: I1203 16:24:02.424366 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Dec 03 16:24:03 crc kubenswrapper[4998]: I1203 16:24:03.251235 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f44dc4fd9-2rfng" event={"ID":"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f","Type":"ContainerStarted","Data":"fda84301c322007bb077cfcbd766a02c487ae1fbbb66774f2ceeddaf55395de0"} Dec 03 16:24:03 crc kubenswrapper[4998]: I1203 16:24:03.258530 4998 generic.go:334] "Generic (PLEG): container finished" podID="e968cb62-c964-46f9-ac6d-6cb674c8bf22" containerID="3b071df7491234fd479ab5af697909c349b263dc051375c3ab6f1c89a97bad20" exitCode=143 Dec 03 16:24:03 crc kubenswrapper[4998]: I1203 16:24:03.258616 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"e968cb62-c964-46f9-ac6d-6cb674c8bf22","Type":"ContainerDied","Data":"3b071df7491234fd479ab5af697909c349b263dc051375c3ab6f1c89a97bad20"} Dec 03 16:24:03 crc kubenswrapper[4998]: I1203 16:24:03.486236 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Dec 03 16:24:03 crc kubenswrapper[4998]: I1203 16:24:03.986962 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="e968cb62-c964-46f9-ac6d-6cb674c8bf22" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.147:9322/\": read tcp 10.217.0.2:49932->10.217.0.147:9322: read: connection reset by peer" Dec 03 16:24:04 crc kubenswrapper[4998]: I1203 16:24:04.288282 4998 generic.go:334] "Generic (PLEG): container finished" podID="e968cb62-c964-46f9-ac6d-6cb674c8bf22" containerID="2e5f63217f2477382afa7cfbebe41754b712ad2a5e18028b07c3c24bd16131f4" exitCode=0 Dec 03 16:24:04 crc kubenswrapper[4998]: I1203 16:24:04.288345 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"e968cb62-c964-46f9-ac6d-6cb674c8bf22","Type":"ContainerDied","Data":"2e5f63217f2477382afa7cfbebe41754b712ad2a5e18028b07c3c24bd16131f4"} Dec 03 16:24:05 crc kubenswrapper[4998]: I1203 16:24:05.301099 4998 generic.go:334] "Generic (PLEG): container finished" podID="aa92f53b-42a5-4b6d-aba4-3bec5917cd59" containerID="dce18792baccb8a737a1d6afcfdeae6a5b72eb876a08b9697f50acde4e2584ba" exitCode=0 Dec 03 16:24:05 crc kubenswrapper[4998]: I1203 16:24:05.301216 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hd2nb" event={"ID":"aa92f53b-42a5-4b6d-aba4-3bec5917cd59","Type":"ContainerDied","Data":"dce18792baccb8a737a1d6afcfdeae6a5b72eb876a08b9697f50acde4e2584ba"} Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.215108 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-67bd586fc-x2g2p"] Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.259884 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7565d78f8d-jm9qt"] Dec 03 16:24:06 crc kubenswrapper[4998]: E1203 16:24:06.260374 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd1aa1ba-fae8-4b45-8af0-a83948b2dd12" containerName="init" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.260400 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd1aa1ba-fae8-4b45-8af0-a83948b2dd12" containerName="init" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.260607 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd1aa1ba-fae8-4b45-8af0-a83948b2dd12" containerName="init" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.261627 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.263910 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.272934 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7565d78f8d-jm9qt"] Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.322650 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f44dc4fd9-2rfng"] Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.340919 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7c5f7f9446-f97x8"] Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.344345 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.359799 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c5f7f9446-f97x8"] Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.414043 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7600639-7175-4c0a-9524-40adb505d3ea-logs\") pod \"horizon-7565d78f8d-jm9qt\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.414093 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7600639-7175-4c0a-9524-40adb505d3ea-horizon-tls-certs\") pod \"horizon-7565d78f8d-jm9qt\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.414128 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d7600639-7175-4c0a-9524-40adb505d3ea-horizon-secret-key\") pod \"horizon-7565d78f8d-jm9qt\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.414170 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24f1f198-c11a-4b5d-bcdd-d359ed88d97f-scripts\") pod \"horizon-7c5f7f9446-f97x8\" (UID: \"24f1f198-c11a-4b5d-bcdd-d359ed88d97f\") " pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.414189 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7600639-7175-4c0a-9524-40adb505d3ea-combined-ca-bundle\") pod \"horizon-7565d78f8d-jm9qt\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.414239 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/24f1f198-c11a-4b5d-bcdd-d359ed88d97f-horizon-tls-certs\") pod \"horizon-7c5f7f9446-f97x8\" (UID: \"24f1f198-c11a-4b5d-bcdd-d359ed88d97f\") " pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.414290 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24f1f198-c11a-4b5d-bcdd-d359ed88d97f-logs\") pod \"horizon-7c5f7f9446-f97x8\" (UID: \"24f1f198-c11a-4b5d-bcdd-d359ed88d97f\") " pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.414338 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/24f1f198-c11a-4b5d-bcdd-d359ed88d97f-horizon-secret-key\") pod \"horizon-7c5f7f9446-f97x8\" (UID: \"24f1f198-c11a-4b5d-bcdd-d359ed88d97f\") " pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.414368 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7600639-7175-4c0a-9524-40adb505d3ea-scripts\") pod \"horizon-7565d78f8d-jm9qt\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.414503 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lt2b\" (UniqueName: \"kubernetes.io/projected/d7600639-7175-4c0a-9524-40adb505d3ea-kube-api-access-2lt2b\") pod \"horizon-7565d78f8d-jm9qt\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.414609 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d7600639-7175-4c0a-9524-40adb505d3ea-config-data\") pod \"horizon-7565d78f8d-jm9qt\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.414710 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f1f198-c11a-4b5d-bcdd-d359ed88d97f-combined-ca-bundle\") pod \"horizon-7c5f7f9446-f97x8\" (UID: \"24f1f198-c11a-4b5d-bcdd-d359ed88d97f\") " pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.414796 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wtb8\" (UniqueName: \"kubernetes.io/projected/24f1f198-c11a-4b5d-bcdd-d359ed88d97f-kube-api-access-4wtb8\") pod \"horizon-7c5f7f9446-f97x8\" (UID: \"24f1f198-c11a-4b5d-bcdd-d359ed88d97f\") " pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.414860 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/24f1f198-c11a-4b5d-bcdd-d359ed88d97f-config-data\") pod \"horizon-7c5f7f9446-f97x8\" (UID: \"24f1f198-c11a-4b5d-bcdd-d359ed88d97f\") " pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.516723 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24f1f198-c11a-4b5d-bcdd-d359ed88d97f-scripts\") pod \"horizon-7c5f7f9446-f97x8\" (UID: \"24f1f198-c11a-4b5d-bcdd-d359ed88d97f\") " pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.517066 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7600639-7175-4c0a-9524-40adb505d3ea-combined-ca-bundle\") pod \"horizon-7565d78f8d-jm9qt\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.517094 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/24f1f198-c11a-4b5d-bcdd-d359ed88d97f-horizon-tls-certs\") pod \"horizon-7c5f7f9446-f97x8\" (UID: \"24f1f198-c11a-4b5d-bcdd-d359ed88d97f\") " pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.517111 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24f1f198-c11a-4b5d-bcdd-d359ed88d97f-logs\") pod \"horizon-7c5f7f9446-f97x8\" (UID: \"24f1f198-c11a-4b5d-bcdd-d359ed88d97f\") " pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.517138 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/24f1f198-c11a-4b5d-bcdd-d359ed88d97f-horizon-secret-key\") pod \"horizon-7c5f7f9446-f97x8\" (UID: \"24f1f198-c11a-4b5d-bcdd-d359ed88d97f\") " pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.517159 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7600639-7175-4c0a-9524-40adb505d3ea-scripts\") pod \"horizon-7565d78f8d-jm9qt\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.517188 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lt2b\" (UniqueName: \"kubernetes.io/projected/d7600639-7175-4c0a-9524-40adb505d3ea-kube-api-access-2lt2b\") pod \"horizon-7565d78f8d-jm9qt\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.517212 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d7600639-7175-4c0a-9524-40adb505d3ea-config-data\") pod \"horizon-7565d78f8d-jm9qt\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.517246 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f1f198-c11a-4b5d-bcdd-d359ed88d97f-combined-ca-bundle\") pod \"horizon-7c5f7f9446-f97x8\" (UID: \"24f1f198-c11a-4b5d-bcdd-d359ed88d97f\") " pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.517275 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wtb8\" (UniqueName: \"kubernetes.io/projected/24f1f198-c11a-4b5d-bcdd-d359ed88d97f-kube-api-access-4wtb8\") pod \"horizon-7c5f7f9446-f97x8\" (UID: \"24f1f198-c11a-4b5d-bcdd-d359ed88d97f\") " pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.517298 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/24f1f198-c11a-4b5d-bcdd-d359ed88d97f-config-data\") pod \"horizon-7c5f7f9446-f97x8\" (UID: \"24f1f198-c11a-4b5d-bcdd-d359ed88d97f\") " pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.517322 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7600639-7175-4c0a-9524-40adb505d3ea-logs\") pod \"horizon-7565d78f8d-jm9qt\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.517343 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7600639-7175-4c0a-9524-40adb505d3ea-horizon-tls-certs\") pod \"horizon-7565d78f8d-jm9qt\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.517369 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d7600639-7175-4c0a-9524-40adb505d3ea-horizon-secret-key\") pod \"horizon-7565d78f8d-jm9qt\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.519622 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24f1f198-c11a-4b5d-bcdd-d359ed88d97f-scripts\") pod \"horizon-7c5f7f9446-f97x8\" (UID: \"24f1f198-c11a-4b5d-bcdd-d359ed88d97f\") " pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.524457 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7600639-7175-4c0a-9524-40adb505d3ea-scripts\") pod \"horizon-7565d78f8d-jm9qt\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.525102 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/24f1f198-c11a-4b5d-bcdd-d359ed88d97f-config-data\") pod \"horizon-7c5f7f9446-f97x8\" (UID: \"24f1f198-c11a-4b5d-bcdd-d359ed88d97f\") " pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.526585 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d7600639-7175-4c0a-9524-40adb505d3ea-config-data\") pod \"horizon-7565d78f8d-jm9qt\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.527027 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d7600639-7175-4c0a-9524-40adb505d3ea-horizon-secret-key\") pod \"horizon-7565d78f8d-jm9qt\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.529367 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7600639-7175-4c0a-9524-40adb505d3ea-horizon-tls-certs\") pod \"horizon-7565d78f8d-jm9qt\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.531245 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/24f1f198-c11a-4b5d-bcdd-d359ed88d97f-horizon-secret-key\") pod \"horizon-7c5f7f9446-f97x8\" (UID: \"24f1f198-c11a-4b5d-bcdd-d359ed88d97f\") " pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.531327 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24f1f198-c11a-4b5d-bcdd-d359ed88d97f-logs\") pod \"horizon-7c5f7f9446-f97x8\" (UID: \"24f1f198-c11a-4b5d-bcdd-d359ed88d97f\") " pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.532383 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7600639-7175-4c0a-9524-40adb505d3ea-combined-ca-bundle\") pod \"horizon-7565d78f8d-jm9qt\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.536630 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lt2b\" (UniqueName: \"kubernetes.io/projected/d7600639-7175-4c0a-9524-40adb505d3ea-kube-api-access-2lt2b\") pod \"horizon-7565d78f8d-jm9qt\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.537458 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24f1f198-c11a-4b5d-bcdd-d359ed88d97f-combined-ca-bundle\") pod \"horizon-7c5f7f9446-f97x8\" (UID: \"24f1f198-c11a-4b5d-bcdd-d359ed88d97f\") " pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.541726 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wtb8\" (UniqueName: \"kubernetes.io/projected/24f1f198-c11a-4b5d-bcdd-d359ed88d97f-kube-api-access-4wtb8\") pod \"horizon-7c5f7f9446-f97x8\" (UID: \"24f1f198-c11a-4b5d-bcdd-d359ed88d97f\") " pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.548590 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7600639-7175-4c0a-9524-40adb505d3ea-logs\") pod \"horizon-7565d78f8d-jm9qt\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.552527 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/24f1f198-c11a-4b5d-bcdd-d359ed88d97f-horizon-tls-certs\") pod \"horizon-7c5f7f9446-f97x8\" (UID: \"24f1f198-c11a-4b5d-bcdd-d359ed88d97f\") " pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.596578 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:06 crc kubenswrapper[4998]: I1203 16:24:06.668276 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.108280 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hd2nb" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.231500 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-fernet-keys\") pod \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\" (UID: \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\") " Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.231904 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-credential-keys\") pod \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\" (UID: \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\") " Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.231946 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-combined-ca-bundle\") pod \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\" (UID: \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\") " Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.231986 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-scripts\") pod \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\" (UID: \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\") " Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.232018 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjmsl\" (UniqueName: \"kubernetes.io/projected/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-kube-api-access-gjmsl\") pod \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\" (UID: \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\") " Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.232065 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-config-data\") pod \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\" (UID: \"aa92f53b-42a5-4b6d-aba4-3bec5917cd59\") " Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.241228 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-kube-api-access-gjmsl" (OuterVolumeSpecName: "kube-api-access-gjmsl") pod "aa92f53b-42a5-4b6d-aba4-3bec5917cd59" (UID: "aa92f53b-42a5-4b6d-aba4-3bec5917cd59"). InnerVolumeSpecName "kube-api-access-gjmsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.247172 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "aa92f53b-42a5-4b6d-aba4-3bec5917cd59" (UID: "aa92f53b-42a5-4b6d-aba4-3bec5917cd59"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.247218 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-scripts" (OuterVolumeSpecName: "scripts") pod "aa92f53b-42a5-4b6d-aba4-3bec5917cd59" (UID: "aa92f53b-42a5-4b6d-aba4-3bec5917cd59"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.249693 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "aa92f53b-42a5-4b6d-aba4-3bec5917cd59" (UID: "aa92f53b-42a5-4b6d-aba4-3bec5917cd59"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.301262 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-config-data" (OuterVolumeSpecName: "config-data") pod "aa92f53b-42a5-4b6d-aba4-3bec5917cd59" (UID: "aa92f53b-42a5-4b6d-aba4-3bec5917cd59"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.330928 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa92f53b-42a5-4b6d-aba4-3bec5917cd59" (UID: "aa92f53b-42a5-4b6d-aba4-3bec5917cd59"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.337064 4998 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.337099 4998 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.337114 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.337126 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.337137 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjmsl\" (UniqueName: \"kubernetes.io/projected/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-kube-api-access-gjmsl\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.337151 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa92f53b-42a5-4b6d-aba4-3bec5917cd59-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.366964 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hd2nb" event={"ID":"aa92f53b-42a5-4b6d-aba4-3bec5917cd59","Type":"ContainerDied","Data":"ba4c8297559225b3f4f9e9050c9473293a0139681e97a465e39cd122e444b222"} Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.367001 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba4c8297559225b3f4f9e9050c9473293a0139681e97a465e39cd122e444b222" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.367046 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hd2nb" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.459092 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-hd2nb"] Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.472415 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-hd2nb"] Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.547215 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-bf2md"] Dec 03 16:24:07 crc kubenswrapper[4998]: E1203 16:24:07.547628 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa92f53b-42a5-4b6d-aba4-3bec5917cd59" containerName="keystone-bootstrap" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.547644 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa92f53b-42a5-4b6d-aba4-3bec5917cd59" containerName="keystone-bootstrap" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.547847 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa92f53b-42a5-4b6d-aba4-3bec5917cd59" containerName="keystone-bootstrap" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.548486 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bf2md" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.551269 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.551316 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.552379 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.552420 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ggfph" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.552777 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.557948 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-bf2md"] Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.647680 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-scripts\") pod \"keystone-bootstrap-bf2md\" (UID: \"751f7a77-ec1d-4529-8762-a48fe772c25b\") " pod="openstack/keystone-bootstrap-bf2md" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.647796 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq4jd\" (UniqueName: \"kubernetes.io/projected/751f7a77-ec1d-4529-8762-a48fe772c25b-kube-api-access-mq4jd\") pod \"keystone-bootstrap-bf2md\" (UID: \"751f7a77-ec1d-4529-8762-a48fe772c25b\") " pod="openstack/keystone-bootstrap-bf2md" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.647820 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-credential-keys\") pod \"keystone-bootstrap-bf2md\" (UID: \"751f7a77-ec1d-4529-8762-a48fe772c25b\") " pod="openstack/keystone-bootstrap-bf2md" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.647850 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-combined-ca-bundle\") pod \"keystone-bootstrap-bf2md\" (UID: \"751f7a77-ec1d-4529-8762-a48fe772c25b\") " pod="openstack/keystone-bootstrap-bf2md" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.647864 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-config-data\") pod \"keystone-bootstrap-bf2md\" (UID: \"751f7a77-ec1d-4529-8762-a48fe772c25b\") " pod="openstack/keystone-bootstrap-bf2md" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.647958 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-fernet-keys\") pod \"keystone-bootstrap-bf2md\" (UID: \"751f7a77-ec1d-4529-8762-a48fe772c25b\") " pod="openstack/keystone-bootstrap-bf2md" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.692936 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa92f53b-42a5-4b6d-aba4-3bec5917cd59" path="/var/lib/kubelet/pods/aa92f53b-42a5-4b6d-aba4-3bec5917cd59/volumes" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.750342 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq4jd\" (UniqueName: \"kubernetes.io/projected/751f7a77-ec1d-4529-8762-a48fe772c25b-kube-api-access-mq4jd\") pod \"keystone-bootstrap-bf2md\" (UID: \"751f7a77-ec1d-4529-8762-a48fe772c25b\") " pod="openstack/keystone-bootstrap-bf2md" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.750387 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-credential-keys\") pod \"keystone-bootstrap-bf2md\" (UID: \"751f7a77-ec1d-4529-8762-a48fe772c25b\") " pod="openstack/keystone-bootstrap-bf2md" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.750423 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-combined-ca-bundle\") pod \"keystone-bootstrap-bf2md\" (UID: \"751f7a77-ec1d-4529-8762-a48fe772c25b\") " pod="openstack/keystone-bootstrap-bf2md" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.750438 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-config-data\") pod \"keystone-bootstrap-bf2md\" (UID: \"751f7a77-ec1d-4529-8762-a48fe772c25b\") " pod="openstack/keystone-bootstrap-bf2md" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.750491 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-fernet-keys\") pod \"keystone-bootstrap-bf2md\" (UID: \"751f7a77-ec1d-4529-8762-a48fe772c25b\") " pod="openstack/keystone-bootstrap-bf2md" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.750536 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-scripts\") pod \"keystone-bootstrap-bf2md\" (UID: \"751f7a77-ec1d-4529-8762-a48fe772c25b\") " pod="openstack/keystone-bootstrap-bf2md" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.754585 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-config-data\") pod \"keystone-bootstrap-bf2md\" (UID: \"751f7a77-ec1d-4529-8762-a48fe772c25b\") " pod="openstack/keystone-bootstrap-bf2md" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.754586 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-scripts\") pod \"keystone-bootstrap-bf2md\" (UID: \"751f7a77-ec1d-4529-8762-a48fe772c25b\") " pod="openstack/keystone-bootstrap-bf2md" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.754909 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-credential-keys\") pod \"keystone-bootstrap-bf2md\" (UID: \"751f7a77-ec1d-4529-8762-a48fe772c25b\") " pod="openstack/keystone-bootstrap-bf2md" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.755698 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-combined-ca-bundle\") pod \"keystone-bootstrap-bf2md\" (UID: \"751f7a77-ec1d-4529-8762-a48fe772c25b\") " pod="openstack/keystone-bootstrap-bf2md" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.763560 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-fernet-keys\") pod \"keystone-bootstrap-bf2md\" (UID: \"751f7a77-ec1d-4529-8762-a48fe772c25b\") " pod="openstack/keystone-bootstrap-bf2md" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.767162 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq4jd\" (UniqueName: \"kubernetes.io/projected/751f7a77-ec1d-4529-8762-a48fe772c25b-kube-api-access-mq4jd\") pod \"keystone-bootstrap-bf2md\" (UID: \"751f7a77-ec1d-4529-8762-a48fe772c25b\") " pod="openstack/keystone-bootstrap-bf2md" Dec 03 16:24:07 crc kubenswrapper[4998]: I1203 16:24:07.890907 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bf2md" Dec 03 16:24:08 crc kubenswrapper[4998]: I1203 16:24:08.043899 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" Dec 03 16:24:08 crc kubenswrapper[4998]: I1203 16:24:08.117313 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-946dbfbcf-7n244"] Dec 03 16:24:08 crc kubenswrapper[4998]: I1203 16:24:08.117539 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-946dbfbcf-7n244" podUID="b5c15d74-1721-40f9-bb22-162a95fd98c1" containerName="dnsmasq-dns" containerID="cri-o://ed81737b4f2a66d75bfdb335123457226be42a4d0163ce2162b676f760450e00" gracePeriod=10 Dec 03 16:24:10 crc kubenswrapper[4998]: I1203 16:24:10.392421 4998 generic.go:334] "Generic (PLEG): container finished" podID="b5c15d74-1721-40f9-bb22-162a95fd98c1" containerID="ed81737b4f2a66d75bfdb335123457226be42a4d0163ce2162b676f760450e00" exitCode=0 Dec 03 16:24:10 crc kubenswrapper[4998]: I1203 16:24:10.392519 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-946dbfbcf-7n244" event={"ID":"b5c15d74-1721-40f9-bb22-162a95fd98c1","Type":"ContainerDied","Data":"ed81737b4f2a66d75bfdb335123457226be42a4d0163ce2162b676f760450e00"} Dec 03 16:24:11 crc kubenswrapper[4998]: I1203 16:24:11.983528 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-946dbfbcf-7n244" podUID="b5c15d74-1721-40f9-bb22-162a95fd98c1" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: connect: connection refused" Dec 03 16:24:12 crc kubenswrapper[4998]: I1203 16:24:12.424711 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="e968cb62-c964-46f9-ac6d-6cb674c8bf22" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.147:9322/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 16:24:15 crc kubenswrapper[4998]: E1203 16:24:15.052856 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.32:5001/podified-master-centos10/openstack-placement-api:watcher_latest" Dec 03 16:24:15 crc kubenswrapper[4998]: E1203 16:24:15.053626 4998 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.32:5001/podified-master-centos10/openstack-placement-api:watcher_latest" Dec 03 16:24:15 crc kubenswrapper[4998]: E1203 16:24:15.053869 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:38.102.83.32:5001/podified-master-centos10/openstack-placement-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mk4d5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-qw92l_openstack(5cca43ea-2244-4f5c-8704-8d91634c0ad6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 16:24:15 crc kubenswrapper[4998]: E1203 16:24:15.055084 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-qw92l" podUID="5cca43ea-2244-4f5c-8704-8d91634c0ad6" Dec 03 16:24:15 crc kubenswrapper[4998]: E1203 16:24:15.447259 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.32:5001/podified-master-centos10/openstack-placement-api:watcher_latest\\\"\"" pod="openstack/placement-db-sync-qw92l" podUID="5cca43ea-2244-4f5c-8704-8d91634c0ad6" Dec 03 16:24:16 crc kubenswrapper[4998]: I1203 16:24:16.984252 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-946dbfbcf-7n244" podUID="b5c15d74-1721-40f9-bb22-162a95fd98c1" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: connect: connection refused" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.244316 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.347910 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e968cb62-c964-46f9-ac6d-6cb674c8bf22-logs\") pod \"e968cb62-c964-46f9-ac6d-6cb674c8bf22\" (UID: \"e968cb62-c964-46f9-ac6d-6cb674c8bf22\") " Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.348001 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e968cb62-c964-46f9-ac6d-6cb674c8bf22-config-data\") pod \"e968cb62-c964-46f9-ac6d-6cb674c8bf22\" (UID: \"e968cb62-c964-46f9-ac6d-6cb674c8bf22\") " Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.348102 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/e968cb62-c964-46f9-ac6d-6cb674c8bf22-custom-prometheus-ca\") pod \"e968cb62-c964-46f9-ac6d-6cb674c8bf22\" (UID: \"e968cb62-c964-46f9-ac6d-6cb674c8bf22\") " Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.348125 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e968cb62-c964-46f9-ac6d-6cb674c8bf22-combined-ca-bundle\") pod \"e968cb62-c964-46f9-ac6d-6cb674c8bf22\" (UID: \"e968cb62-c964-46f9-ac6d-6cb674c8bf22\") " Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.348221 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbz7r\" (UniqueName: \"kubernetes.io/projected/e968cb62-c964-46f9-ac6d-6cb674c8bf22-kube-api-access-rbz7r\") pod \"e968cb62-c964-46f9-ac6d-6cb674c8bf22\" (UID: \"e968cb62-c964-46f9-ac6d-6cb674c8bf22\") " Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.348438 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e968cb62-c964-46f9-ac6d-6cb674c8bf22-logs" (OuterVolumeSpecName: "logs") pod "e968cb62-c964-46f9-ac6d-6cb674c8bf22" (UID: "e968cb62-c964-46f9-ac6d-6cb674c8bf22"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.348632 4998 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e968cb62-c964-46f9-ac6d-6cb674c8bf22-logs\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.378883 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e968cb62-c964-46f9-ac6d-6cb674c8bf22-kube-api-access-rbz7r" (OuterVolumeSpecName: "kube-api-access-rbz7r") pod "e968cb62-c964-46f9-ac6d-6cb674c8bf22" (UID: "e968cb62-c964-46f9-ac6d-6cb674c8bf22"). InnerVolumeSpecName "kube-api-access-rbz7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.387563 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e968cb62-c964-46f9-ac6d-6cb674c8bf22-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "e968cb62-c964-46f9-ac6d-6cb674c8bf22" (UID: "e968cb62-c964-46f9-ac6d-6cb674c8bf22"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.401620 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e968cb62-c964-46f9-ac6d-6cb674c8bf22-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e968cb62-c964-46f9-ac6d-6cb674c8bf22" (UID: "e968cb62-c964-46f9-ac6d-6cb674c8bf22"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.425178 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="e968cb62-c964-46f9-ac6d-6cb674c8bf22" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.147:9322/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.425292 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.439814 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e968cb62-c964-46f9-ac6d-6cb674c8bf22-config-data" (OuterVolumeSpecName: "config-data") pod "e968cb62-c964-46f9-ac6d-6cb674c8bf22" (UID: "e968cb62-c964-46f9-ac6d-6cb674c8bf22"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.450613 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbz7r\" (UniqueName: \"kubernetes.io/projected/e968cb62-c964-46f9-ac6d-6cb674c8bf22-kube-api-access-rbz7r\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.450641 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e968cb62-c964-46f9-ac6d-6cb674c8bf22-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.450650 4998 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/e968cb62-c964-46f9-ac6d-6cb674c8bf22-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.450659 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e968cb62-c964-46f9-ac6d-6cb674c8bf22-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.467811 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"e968cb62-c964-46f9-ac6d-6cb674c8bf22","Type":"ContainerDied","Data":"fb0d0504f1fffae1f15467db418cfa442ac6dfcd08ea82fc04ccd61368a5ae9c"} Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.467866 4998 scope.go:117] "RemoveContainer" containerID="2e5f63217f2477382afa7cfbebe41754b712ad2a5e18028b07c3c24bd16131f4" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.467864 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.541953 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.559237 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-api-0"] Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.578368 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Dec 03 16:24:17 crc kubenswrapper[4998]: E1203 16:24:17.578774 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e968cb62-c964-46f9-ac6d-6cb674c8bf22" containerName="watcher-api-log" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.578791 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="e968cb62-c964-46f9-ac6d-6cb674c8bf22" containerName="watcher-api-log" Dec 03 16:24:17 crc kubenswrapper[4998]: E1203 16:24:17.578822 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e968cb62-c964-46f9-ac6d-6cb674c8bf22" containerName="watcher-api" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.578830 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="e968cb62-c964-46f9-ac6d-6cb674c8bf22" containerName="watcher-api" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.579008 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="e968cb62-c964-46f9-ac6d-6cb674c8bf22" containerName="watcher-api-log" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.579025 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="e968cb62-c964-46f9-ac6d-6cb674c8bf22" containerName="watcher-api" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.580118 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.583873 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.595993 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.655261 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41e16c37-8ac4-40c7-8873-fcb8364f31dc-logs\") pod \"watcher-api-0\" (UID: \"41e16c37-8ac4-40c7-8873-fcb8364f31dc\") " pod="openstack/watcher-api-0" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.655428 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41e16c37-8ac4-40c7-8873-fcb8364f31dc-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"41e16c37-8ac4-40c7-8873-fcb8364f31dc\") " pod="openstack/watcher-api-0" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.655594 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/41e16c37-8ac4-40c7-8873-fcb8364f31dc-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"41e16c37-8ac4-40c7-8873-fcb8364f31dc\") " pod="openstack/watcher-api-0" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.655823 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hx79k\" (UniqueName: \"kubernetes.io/projected/41e16c37-8ac4-40c7-8873-fcb8364f31dc-kube-api-access-hx79k\") pod \"watcher-api-0\" (UID: \"41e16c37-8ac4-40c7-8873-fcb8364f31dc\") " pod="openstack/watcher-api-0" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.655914 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41e16c37-8ac4-40c7-8873-fcb8364f31dc-config-data\") pod \"watcher-api-0\" (UID: \"41e16c37-8ac4-40c7-8873-fcb8364f31dc\") " pod="openstack/watcher-api-0" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.688650 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e968cb62-c964-46f9-ac6d-6cb674c8bf22" path="/var/lib/kubelet/pods/e968cb62-c964-46f9-ac6d-6cb674c8bf22/volumes" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.757578 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hx79k\" (UniqueName: \"kubernetes.io/projected/41e16c37-8ac4-40c7-8873-fcb8364f31dc-kube-api-access-hx79k\") pod \"watcher-api-0\" (UID: \"41e16c37-8ac4-40c7-8873-fcb8364f31dc\") " pod="openstack/watcher-api-0" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.757664 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41e16c37-8ac4-40c7-8873-fcb8364f31dc-config-data\") pod \"watcher-api-0\" (UID: \"41e16c37-8ac4-40c7-8873-fcb8364f31dc\") " pod="openstack/watcher-api-0" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.757727 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41e16c37-8ac4-40c7-8873-fcb8364f31dc-logs\") pod \"watcher-api-0\" (UID: \"41e16c37-8ac4-40c7-8873-fcb8364f31dc\") " pod="openstack/watcher-api-0" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.757824 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41e16c37-8ac4-40c7-8873-fcb8364f31dc-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"41e16c37-8ac4-40c7-8873-fcb8364f31dc\") " pod="openstack/watcher-api-0" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.757974 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/41e16c37-8ac4-40c7-8873-fcb8364f31dc-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"41e16c37-8ac4-40c7-8873-fcb8364f31dc\") " pod="openstack/watcher-api-0" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.758251 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41e16c37-8ac4-40c7-8873-fcb8364f31dc-logs\") pod \"watcher-api-0\" (UID: \"41e16c37-8ac4-40c7-8873-fcb8364f31dc\") " pod="openstack/watcher-api-0" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.764063 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41e16c37-8ac4-40c7-8873-fcb8364f31dc-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"41e16c37-8ac4-40c7-8873-fcb8364f31dc\") " pod="openstack/watcher-api-0" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.764087 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/41e16c37-8ac4-40c7-8873-fcb8364f31dc-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"41e16c37-8ac4-40c7-8873-fcb8364f31dc\") " pod="openstack/watcher-api-0" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.764326 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41e16c37-8ac4-40c7-8873-fcb8364f31dc-config-data\") pod \"watcher-api-0\" (UID: \"41e16c37-8ac4-40c7-8873-fcb8364f31dc\") " pod="openstack/watcher-api-0" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.780638 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hx79k\" (UniqueName: \"kubernetes.io/projected/41e16c37-8ac4-40c7-8873-fcb8364f31dc-kube-api-access-hx79k\") pod \"watcher-api-0\" (UID: \"41e16c37-8ac4-40c7-8873-fcb8364f31dc\") " pod="openstack/watcher-api-0" Dec 03 16:24:17 crc kubenswrapper[4998]: I1203 16:24:17.914337 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 03 16:24:26 crc kubenswrapper[4998]: I1203 16:24:26.983173 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-946dbfbcf-7n244" podUID="b5c15d74-1721-40f9-bb22-162a95fd98c1" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: i/o timeout" Dec 03 16:24:26 crc kubenswrapper[4998]: I1203 16:24:26.985129 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-946dbfbcf-7n244" Dec 03 16:24:27 crc kubenswrapper[4998]: I1203 16:24:27.111957 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:24:27 crc kubenswrapper[4998]: I1203 16:24:27.112229 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:24:29 crc kubenswrapper[4998]: E1203 16:24:29.968340 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.32:5001/podified-master-centos10/openstack-horizon:watcher_latest" Dec 03 16:24:29 crc kubenswrapper[4998]: E1203 16:24:29.968667 4998 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.32:5001/podified-master-centos10/openstack-horizon:watcher_latest" Dec 03 16:24:29 crc kubenswrapper[4998]: E1203 16:24:29.968849 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:38.102.83.32:5001/podified-master-centos10/openstack-horizon:watcher_latest,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n658h5fh668h9fh5cdh668h5c7h5cch86h66h67dhf7h579h5b4hch5ddh57dh645h545h6dh8h55fh656h5b8h56fhd8h5f9h5b6h599h56ch5f6h59cq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lwlcw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7f44dc4fd9-2rfng_openstack(6797c9e9-d8fb-4c61-8343-8bc2742fdc1f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 16:24:29 crc kubenswrapper[4998]: E1203 16:24:29.980319 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.32:5001/podified-master-centos10/openstack-horizon:watcher_latest\\\"\"]" pod="openstack/horizon-7f44dc4fd9-2rfng" podUID="6797c9e9-d8fb-4c61-8343-8bc2742fdc1f" Dec 03 16:24:31 crc kubenswrapper[4998]: E1203 16:24:31.297380 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.32:5001/podified-master-centos10/openstack-glance-api:watcher_latest" Dec 03 16:24:31 crc kubenswrapper[4998]: E1203 16:24:31.298010 4998 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.32:5001/podified-master-centos10/openstack-glance-api:watcher_latest" Dec 03 16:24:31 crc kubenswrapper[4998]: E1203 16:24:31.298146 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:38.102.83.32:5001/podified-master-centos10/openstack-glance-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ftjd7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-2rqhm_openstack(de77cff6-7039-418c-8114-3c64ea4ce8e4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 16:24:31 crc kubenswrapper[4998]: E1203 16:24:31.299301 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-2rqhm" podUID="de77cff6-7039-418c-8114-3c64ea4ce8e4" Dec 03 16:24:31 crc kubenswrapper[4998]: E1203 16:24:31.318981 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.32:5001/podified-master-centos10/openstack-barbican-api:watcher_latest" Dec 03 16:24:31 crc kubenswrapper[4998]: E1203 16:24:31.319033 4998 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.32:5001/podified-master-centos10/openstack-barbican-api:watcher_latest" Dec 03 16:24:31 crc kubenswrapper[4998]: E1203 16:24:31.319176 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:38.102.83.32:5001/podified-master-centos10/openstack-barbican-api:watcher_latest,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pfq6t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-w8k5f_openstack(5d2429d8-9baa-4532-9f8b-9ffb72a3a764): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 16:24:31 crc kubenswrapper[4998]: E1203 16:24:31.320876 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-w8k5f" podUID="5d2429d8-9baa-4532-9f8b-9ffb72a3a764" Dec 03 16:24:31 crc kubenswrapper[4998]: I1203 16:24:31.445040 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-946dbfbcf-7n244" Dec 03 16:24:31 crc kubenswrapper[4998]: I1203 16:24:31.570808 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-ovsdbserver-nb\") pod \"b5c15d74-1721-40f9-bb22-162a95fd98c1\" (UID: \"b5c15d74-1721-40f9-bb22-162a95fd98c1\") " Dec 03 16:24:31 crc kubenswrapper[4998]: I1203 16:24:31.570979 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-ovsdbserver-sb\") pod \"b5c15d74-1721-40f9-bb22-162a95fd98c1\" (UID: \"b5c15d74-1721-40f9-bb22-162a95fd98c1\") " Dec 03 16:24:31 crc kubenswrapper[4998]: I1203 16:24:31.571086 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-config\") pod \"b5c15d74-1721-40f9-bb22-162a95fd98c1\" (UID: \"b5c15d74-1721-40f9-bb22-162a95fd98c1\") " Dec 03 16:24:31 crc kubenswrapper[4998]: I1203 16:24:31.571119 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vlnw\" (UniqueName: \"kubernetes.io/projected/b5c15d74-1721-40f9-bb22-162a95fd98c1-kube-api-access-8vlnw\") pod \"b5c15d74-1721-40f9-bb22-162a95fd98c1\" (UID: \"b5c15d74-1721-40f9-bb22-162a95fd98c1\") " Dec 03 16:24:31 crc kubenswrapper[4998]: I1203 16:24:31.571203 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-dns-svc\") pod \"b5c15d74-1721-40f9-bb22-162a95fd98c1\" (UID: \"b5c15d74-1721-40f9-bb22-162a95fd98c1\") " Dec 03 16:24:31 crc kubenswrapper[4998]: I1203 16:24:31.571249 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-dns-swift-storage-0\") pod \"b5c15d74-1721-40f9-bb22-162a95fd98c1\" (UID: \"b5c15d74-1721-40f9-bb22-162a95fd98c1\") " Dec 03 16:24:31 crc kubenswrapper[4998]: I1203 16:24:31.582971 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5c15d74-1721-40f9-bb22-162a95fd98c1-kube-api-access-8vlnw" (OuterVolumeSpecName: "kube-api-access-8vlnw") pod "b5c15d74-1721-40f9-bb22-162a95fd98c1" (UID: "b5c15d74-1721-40f9-bb22-162a95fd98c1"). InnerVolumeSpecName "kube-api-access-8vlnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:24:31 crc kubenswrapper[4998]: I1203 16:24:31.620931 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-946dbfbcf-7n244" event={"ID":"b5c15d74-1721-40f9-bb22-162a95fd98c1","Type":"ContainerDied","Data":"5d41092474acf4cc286d3ab50aeae4bac75de306e3f04da4736a2d5a7fd2540c"} Dec 03 16:24:31 crc kubenswrapper[4998]: I1203 16:24:31.620952 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-946dbfbcf-7n244" Dec 03 16:24:31 crc kubenswrapper[4998]: E1203 16:24:31.623519 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.32:5001/podified-master-centos10/openstack-barbican-api:watcher_latest\\\"\"" pod="openstack/barbican-db-sync-w8k5f" podUID="5d2429d8-9baa-4532-9f8b-9ffb72a3a764" Dec 03 16:24:31 crc kubenswrapper[4998]: I1203 16:24:31.624937 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b5c15d74-1721-40f9-bb22-162a95fd98c1" (UID: "b5c15d74-1721-40f9-bb22-162a95fd98c1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:24:31 crc kubenswrapper[4998]: I1203 16:24:31.626204 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b5c15d74-1721-40f9-bb22-162a95fd98c1" (UID: "b5c15d74-1721-40f9-bb22-162a95fd98c1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:24:31 crc kubenswrapper[4998]: I1203 16:24:31.639346 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b5c15d74-1721-40f9-bb22-162a95fd98c1" (UID: "b5c15d74-1721-40f9-bb22-162a95fd98c1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:24:31 crc kubenswrapper[4998]: I1203 16:24:31.653169 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b5c15d74-1721-40f9-bb22-162a95fd98c1" (UID: "b5c15d74-1721-40f9-bb22-162a95fd98c1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:24:31 crc kubenswrapper[4998]: I1203 16:24:31.659456 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-config" (OuterVolumeSpecName: "config") pod "b5c15d74-1721-40f9-bb22-162a95fd98c1" (UID: "b5c15d74-1721-40f9-bb22-162a95fd98c1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:24:31 crc kubenswrapper[4998]: I1203 16:24:31.673199 4998 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:31 crc kubenswrapper[4998]: I1203 16:24:31.673243 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:31 crc kubenswrapper[4998]: I1203 16:24:31.673260 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vlnw\" (UniqueName: \"kubernetes.io/projected/b5c15d74-1721-40f9-bb22-162a95fd98c1-kube-api-access-8vlnw\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:31 crc kubenswrapper[4998]: I1203 16:24:31.673274 4998 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:31 crc kubenswrapper[4998]: I1203 16:24:31.673284 4998 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:31 crc kubenswrapper[4998]: I1203 16:24:31.673294 4998 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5c15d74-1721-40f9-bb22-162a95fd98c1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:31 crc kubenswrapper[4998]: I1203 16:24:31.948054 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-946dbfbcf-7n244"] Dec 03 16:24:31 crc kubenswrapper[4998]: I1203 16:24:31.957790 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-946dbfbcf-7n244"] Dec 03 16:24:31 crc kubenswrapper[4998]: I1203 16:24:31.985549 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-946dbfbcf-7n244" podUID="b5c15d74-1721-40f9-bb22-162a95fd98c1" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: i/o timeout" Dec 03 16:24:32 crc kubenswrapper[4998]: I1203 16:24:32.882414 4998 scope.go:117] "RemoveContainer" containerID="3b071df7491234fd479ab5af697909c349b263dc051375c3ab6f1c89a97bad20" Dec 03 16:24:32 crc kubenswrapper[4998]: E1203 16:24:32.904035 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.32:5001/podified-master-centos10/openstack-cinder-api:watcher_latest" Dec 03 16:24:32 crc kubenswrapper[4998]: E1203 16:24:32.904092 4998 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.32:5001/podified-master-centos10/openstack-cinder-api:watcher_latest" Dec 03 16:24:32 crc kubenswrapper[4998]: E1203 16:24:32.904231 4998 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:38.102.83.32:5001/podified-master-centos10/openstack-cinder-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jnlxq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-8lz9j_openstack(82aa9d23-457c-45a0-9729-e5a803738237): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 16:24:32 crc kubenswrapper[4998]: E1203 16:24:32.905702 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-8lz9j" podUID="82aa9d23-457c-45a0-9729-e5a803738237" Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.026044 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f44dc4fd9-2rfng" Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.202223 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-config-data\") pod \"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f\" (UID: \"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f\") " Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.202285 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-scripts\") pod \"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f\" (UID: \"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f\") " Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.202415 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-horizon-secret-key\") pod \"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f\" (UID: \"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f\") " Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.202449 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-logs\") pod \"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f\" (UID: \"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f\") " Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.202512 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwlcw\" (UniqueName: \"kubernetes.io/projected/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-kube-api-access-lwlcw\") pod \"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f\" (UID: \"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f\") " Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.203189 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-scripts" (OuterVolumeSpecName: "scripts") pod "6797c9e9-d8fb-4c61-8343-8bc2742fdc1f" (UID: "6797c9e9-d8fb-4c61-8343-8bc2742fdc1f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.203464 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-config-data" (OuterVolumeSpecName: "config-data") pod "6797c9e9-d8fb-4c61-8343-8bc2742fdc1f" (UID: "6797c9e9-d8fb-4c61-8343-8bc2742fdc1f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.204152 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-logs" (OuterVolumeSpecName: "logs") pod "6797c9e9-d8fb-4c61-8343-8bc2742fdc1f" (UID: "6797c9e9-d8fb-4c61-8343-8bc2742fdc1f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.208285 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-kube-api-access-lwlcw" (OuterVolumeSpecName: "kube-api-access-lwlcw") pod "6797c9e9-d8fb-4c61-8343-8bc2742fdc1f" (UID: "6797c9e9-d8fb-4c61-8343-8bc2742fdc1f"). InnerVolumeSpecName "kube-api-access-lwlcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.211838 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "6797c9e9-d8fb-4c61-8343-8bc2742fdc1f" (UID: "6797c9e9-d8fb-4c61-8343-8bc2742fdc1f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.304398 4998 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.304422 4998 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-logs\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.304431 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwlcw\" (UniqueName: \"kubernetes.io/projected/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-kube-api-access-lwlcw\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.304439 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.304448 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.364657 4998 scope.go:117] "RemoveContainer" containerID="ed81737b4f2a66d75bfdb335123457226be42a4d0163ce2162b676f760450e00" Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.383046 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7565d78f8d-jm9qt"] Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.405032 4998 scope.go:117] "RemoveContainer" containerID="c3e6d8b3107f8f81a007fb86a9f63de459dd96469026a585ee40be984abc2b64" Dec 03 16:24:33 crc kubenswrapper[4998]: W1203 16:24:33.412386 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7600639_7175_4c0a_9524_40adb505d3ea.slice/crio-6eb3a39450e43793101cc3b0a6e6381ea10674fde4b0cc6eedd34b0de43e1c1b WatchSource:0}: Error finding container 6eb3a39450e43793101cc3b0a6e6381ea10674fde4b0cc6eedd34b0de43e1c1b: Status 404 returned error can't find the container with id 6eb3a39450e43793101cc3b0a6e6381ea10674fde4b0cc6eedd34b0de43e1c1b Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.638061 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.649470 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-bf2md"] Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.653097 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7565d78f8d-jm9qt" event={"ID":"d7600639-7175-4c0a-9524-40adb505d3ea","Type":"ContainerStarted","Data":"6eb3a39450e43793101cc3b0a6e6381ea10674fde4b0cc6eedd34b0de43e1c1b"} Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.657859 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c5f7f9446-f97x8"] Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.661346 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4faa24a-a43f-4d24-b20d-09e241d0cc96","Type":"ContainerStarted","Data":"5ae86f8014172d9b95a4494ff98b6cbc3dfc4a7399092fa9f3e71c9260a39e54"} Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.706943 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f44dc4fd9-2rfng" Dec 03 16:24:33 crc kubenswrapper[4998]: E1203 16:24:33.716981 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.32:5001/podified-master-centos10/openstack-cinder-api:watcher_latest\\\"\"" pod="openstack/cinder-db-sync-8lz9j" podUID="82aa9d23-457c-45a0-9729-e5a803738237" Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.719957 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5c15d74-1721-40f9-bb22-162a95fd98c1" path="/var/lib/kubelet/pods/b5c15d74-1721-40f9-bb22-162a95fd98c1/volumes" Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.720933 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-648fb89557-mvw59" event={"ID":"53f87c55-ecee-4d26-8244-6181b21dece0","Type":"ContainerStarted","Data":"95037244abd2cb3d3c77b85237199f52984305069d5d070b3c10e0a4cdefeab5"} Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.720966 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"8e3693fc-394d-4cbb-8501-34bbc0483c37","Type":"ContainerStarted","Data":"7de1267e2d008561febdc4f064e793bf1f979aae3c60266323b9c300d74fd5bd"} Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.720982 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f44dc4fd9-2rfng" event={"ID":"6797c9e9-d8fb-4c61-8343-8bc2742fdc1f","Type":"ContainerDied","Data":"fda84301c322007bb077cfcbd766a02c487ae1fbbb66774f2ceeddaf55395de0"} Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.734251 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=4.371598469 podStartE2EDuration="36.734230379s" podCreationTimestamp="2025-12-03 16:23:57 +0000 UTC" firstStartedPulling="2025-12-03 16:23:58.952975425 +0000 UTC m=+1217.564675648" lastFinishedPulling="2025-12-03 16:24:31.315607315 +0000 UTC m=+1249.927307558" observedRunningTime="2025-12-03 16:24:33.711452635 +0000 UTC m=+1252.323152888" watchObservedRunningTime="2025-12-03 16:24:33.734230379 +0000 UTC m=+1252.345930612" Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.798363 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f44dc4fd9-2rfng"] Dec 03 16:24:33 crc kubenswrapper[4998]: I1203 16:24:33.809631 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7f44dc4fd9-2rfng"] Dec 03 16:24:34 crc kubenswrapper[4998]: I1203 16:24:34.718092 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"41e16c37-8ac4-40c7-8873-fcb8364f31dc","Type":"ContainerStarted","Data":"9035151ef65f166074aafe72a2fa23bdf67f54d4200dd95dc8aa8199cee7ca2b"} Dec 03 16:24:34 crc kubenswrapper[4998]: I1203 16:24:34.718628 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"41e16c37-8ac4-40c7-8873-fcb8364f31dc","Type":"ContainerStarted","Data":"e134225c8bcc72bb0f67575ff29b1c2c42310ace7087102fd3361dc42d90d7b6"} Dec 03 16:24:34 crc kubenswrapper[4998]: I1203 16:24:34.723888 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c5f7f9446-f97x8" event={"ID":"24f1f198-c11a-4b5d-bcdd-d359ed88d97f","Type":"ContainerStarted","Data":"a44e8f0c8ef8a12ad86e78a48a2e9cf9722581de4bfe57e467fe99f3d31ecd75"} Dec 03 16:24:34 crc kubenswrapper[4998]: I1203 16:24:34.723941 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c5f7f9446-f97x8" event={"ID":"24f1f198-c11a-4b5d-bcdd-d359ed88d97f","Type":"ContainerStarted","Data":"6897db2d7b3c29a75b38685d8137a81eb69e1341926265ba4c8da76d392d7c7e"} Dec 03 16:24:34 crc kubenswrapper[4998]: I1203 16:24:34.725880 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67bd586fc-x2g2p" event={"ID":"ef216407-b6ec-4691-8f49-cf8fce3f3544","Type":"ContainerStarted","Data":"d4b7cb504608696bdfe943b2476e64150c5012ce77fbcbf7167ba50a7e549607"} Dec 03 16:24:34 crc kubenswrapper[4998]: I1203 16:24:34.725913 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67bd586fc-x2g2p" event={"ID":"ef216407-b6ec-4691-8f49-cf8fce3f3544","Type":"ContainerStarted","Data":"acfefbd1e9cf3e1a042744e581a4aa19f86d8779dde8da25311f8659c3f3bd73"} Dec 03 16:24:34 crc kubenswrapper[4998]: I1203 16:24:34.726052 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-67bd586fc-x2g2p" podUID="ef216407-b6ec-4691-8f49-cf8fce3f3544" containerName="horizon-log" containerID="cri-o://acfefbd1e9cf3e1a042744e581a4aa19f86d8779dde8da25311f8659c3f3bd73" gracePeriod=30 Dec 03 16:24:34 crc kubenswrapper[4998]: I1203 16:24:34.726734 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-67bd586fc-x2g2p" podUID="ef216407-b6ec-4691-8f49-cf8fce3f3544" containerName="horizon" containerID="cri-o://d4b7cb504608696bdfe943b2476e64150c5012ce77fbcbf7167ba50a7e549607" gracePeriod=30 Dec 03 16:24:34 crc kubenswrapper[4998]: I1203 16:24:34.729217 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"5b94f565-0948-4e4f-9636-2833d43bb829","Type":"ContainerStarted","Data":"2c4f81d00299bd0d05b7f51e1c0c848a779a1b0b6f199befff2e205150316838"} Dec 03 16:24:34 crc kubenswrapper[4998]: I1203 16:24:34.733512 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-648fb89557-mvw59" event={"ID":"53f87c55-ecee-4d26-8244-6181b21dece0","Type":"ContainerStarted","Data":"72038f1b94d45b72f211d0025db98351b1b4c797dfa770ea034904cf1c1b58ff"} Dec 03 16:24:34 crc kubenswrapper[4998]: I1203 16:24:34.733680 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-648fb89557-mvw59" podUID="53f87c55-ecee-4d26-8244-6181b21dece0" containerName="horizon-log" containerID="cri-o://95037244abd2cb3d3c77b85237199f52984305069d5d070b3c10e0a4cdefeab5" gracePeriod=30 Dec 03 16:24:34 crc kubenswrapper[4998]: I1203 16:24:34.733831 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-648fb89557-mvw59" podUID="53f87c55-ecee-4d26-8244-6181b21dece0" containerName="horizon" containerID="cri-o://72038f1b94d45b72f211d0025db98351b1b4c797dfa770ea034904cf1c1b58ff" gracePeriod=30 Dec 03 16:24:34 crc kubenswrapper[4998]: I1203 16:24:34.742170 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bf2md" event={"ID":"751f7a77-ec1d-4529-8762-a48fe772c25b","Type":"ContainerStarted","Data":"3e0b15332e4f017b612e36e98b7d4c7901623e0813d95832b81628ee8a6bbaeb"} Dec 03 16:24:34 crc kubenswrapper[4998]: I1203 16:24:34.742225 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bf2md" event={"ID":"751f7a77-ec1d-4529-8762-a48fe772c25b","Type":"ContainerStarted","Data":"b0b836f41037c77bb86a6126a532b972f0cc30ed09612bcdc243c389f93c792f"} Dec 03 16:24:34 crc kubenswrapper[4998]: I1203 16:24:34.744786 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qw92l" event={"ID":"5cca43ea-2244-4f5c-8704-8d91634c0ad6","Type":"ContainerStarted","Data":"a2dafe752937450125658f1ce9546da3c4d740014f80b43ed2fba0e66e8931f8"} Dec 03 16:24:34 crc kubenswrapper[4998]: I1203 16:24:34.752277 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-67bd586fc-x2g2p" podStartSLOduration=3.880906323 podStartE2EDuration="37.752257852s" podCreationTimestamp="2025-12-03 16:23:57 +0000 UTC" firstStartedPulling="2025-12-03 16:23:59.047996698 +0000 UTC m=+1217.659696911" lastFinishedPulling="2025-12-03 16:24:32.919348197 +0000 UTC m=+1251.531048440" observedRunningTime="2025-12-03 16:24:34.745873017 +0000 UTC m=+1253.357573260" watchObservedRunningTime="2025-12-03 16:24:34.752257852 +0000 UTC m=+1253.363958075" Dec 03 16:24:34 crc kubenswrapper[4998]: I1203 16:24:34.760662 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7565d78f8d-jm9qt" event={"ID":"d7600639-7175-4c0a-9524-40adb505d3ea","Type":"ContainerStarted","Data":"549d68f35b182d6fa1f0327a26451914c3855dd97314b4621e055785c688e514"} Dec 03 16:24:34 crc kubenswrapper[4998]: I1203 16:24:34.760726 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7565d78f8d-jm9qt" event={"ID":"d7600639-7175-4c0a-9524-40adb505d3ea","Type":"ContainerStarted","Data":"75a61917c7bd2bcb544cb164de1c7e5b8e8e2cbe5262c7fa8075c0db387aa835"} Dec 03 16:24:34 crc kubenswrapper[4998]: I1203 16:24:34.789625 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-bf2md" podStartSLOduration=27.789605231 podStartE2EDuration="27.789605231s" podCreationTimestamp="2025-12-03 16:24:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:24:34.768966099 +0000 UTC m=+1253.380666332" watchObservedRunningTime="2025-12-03 16:24:34.789605231 +0000 UTC m=+1253.401305454" Dec 03 16:24:34 crc kubenswrapper[4998]: I1203 16:24:34.796554 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-qw92l" podStartSLOduration=3.613357434 podStartE2EDuration="37.79653779s" podCreationTimestamp="2025-12-03 16:23:57 +0000 UTC" firstStartedPulling="2025-12-03 16:23:59.229256975 +0000 UTC m=+1217.840957198" lastFinishedPulling="2025-12-03 16:24:33.412437331 +0000 UTC m=+1252.024137554" observedRunningTime="2025-12-03 16:24:34.786056545 +0000 UTC m=+1253.397756778" watchObservedRunningTime="2025-12-03 16:24:34.79653779 +0000 UTC m=+1253.408238013" Dec 03 16:24:34 crc kubenswrapper[4998]: I1203 16:24:34.810575 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-applier-0" podStartSLOduration=5.037548081 podStartE2EDuration="37.81054692s" podCreationTimestamp="2025-12-03 16:23:57 +0000 UTC" firstStartedPulling="2025-12-03 16:23:58.542649127 +0000 UTC m=+1217.154349350" lastFinishedPulling="2025-12-03 16:24:31.315647966 +0000 UTC m=+1249.927348189" observedRunningTime="2025-12-03 16:24:34.804833101 +0000 UTC m=+1253.416533324" watchObservedRunningTime="2025-12-03 16:24:34.81054692 +0000 UTC m=+1253.422247153" Dec 03 16:24:34 crc kubenswrapper[4998]: I1203 16:24:34.827720 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-648fb89557-mvw59" podStartSLOduration=3.945647803 podStartE2EDuration="37.827700108s" podCreationTimestamp="2025-12-03 16:23:57 +0000 UTC" firstStartedPulling="2025-12-03 16:23:58.948480621 +0000 UTC m=+1217.560180844" lastFinishedPulling="2025-12-03 16:24:32.830532916 +0000 UTC m=+1251.442233149" observedRunningTime="2025-12-03 16:24:34.826830287 +0000 UTC m=+1253.438530520" watchObservedRunningTime="2025-12-03 16:24:34.827700108 +0000 UTC m=+1253.439400331" Dec 03 16:24:34 crc kubenswrapper[4998]: I1203 16:24:34.854702 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7565d78f8d-jm9qt" podStartSLOduration=28.854686674 podStartE2EDuration="28.854686674s" podCreationTimestamp="2025-12-03 16:24:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:24:34.847001457 +0000 UTC m=+1253.458701680" watchObservedRunningTime="2025-12-03 16:24:34.854686674 +0000 UTC m=+1253.466386887" Dec 03 16:24:35 crc kubenswrapper[4998]: I1203 16:24:35.690906 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6797c9e9-d8fb-4c61-8343-8bc2742fdc1f" path="/var/lib/kubelet/pods/6797c9e9-d8fb-4c61-8343-8bc2742fdc1f/volumes" Dec 03 16:24:35 crc kubenswrapper[4998]: I1203 16:24:35.769776 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c5f7f9446-f97x8" event={"ID":"24f1f198-c11a-4b5d-bcdd-d359ed88d97f","Type":"ContainerStarted","Data":"b8cdabb7f147a31cf2d334782bc99623763fc12bf9ff48af8303edbb4f8ff7ab"} Dec 03 16:24:35 crc kubenswrapper[4998]: I1203 16:24:35.772923 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"41e16c37-8ac4-40c7-8873-fcb8364f31dc","Type":"ContainerStarted","Data":"528c23fd97116c85c9f3479ca1d05476298c5e14f0628192a47e3177573979c4"} Dec 03 16:24:35 crc kubenswrapper[4998]: I1203 16:24:35.773330 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Dec 03 16:24:35 crc kubenswrapper[4998]: I1203 16:24:35.795434 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7c5f7f9446-f97x8" podStartSLOduration=29.795405318 podStartE2EDuration="29.795405318s" podCreationTimestamp="2025-12-03 16:24:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:24:35.791252867 +0000 UTC m=+1254.402953090" watchObservedRunningTime="2025-12-03 16:24:35.795405318 +0000 UTC m=+1254.407105561" Dec 03 16:24:35 crc kubenswrapper[4998]: I1203 16:24:35.826345 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=18.82632357 podStartE2EDuration="18.82632357s" podCreationTimestamp="2025-12-03 16:24:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:24:35.82018011 +0000 UTC m=+1254.431880333" watchObservedRunningTime="2025-12-03 16:24:35.82632357 +0000 UTC m=+1254.438023803" Dec 03 16:24:36 crc kubenswrapper[4998]: I1203 16:24:36.598125 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:36 crc kubenswrapper[4998]: I1203 16:24:36.598204 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:36 crc kubenswrapper[4998]: I1203 16:24:36.669035 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:36 crc kubenswrapper[4998]: I1203 16:24:36.669154 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:37 crc kubenswrapper[4998]: I1203 16:24:37.463324 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-applier-0" Dec 03 16:24:37 crc kubenswrapper[4998]: I1203 16:24:37.463376 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-applier-0" Dec 03 16:24:37 crc kubenswrapper[4998]: I1203 16:24:37.507820 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-applier-0" Dec 03 16:24:37 crc kubenswrapper[4998]: I1203 16:24:37.558211 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-648fb89557-mvw59" Dec 03 16:24:37 crc kubenswrapper[4998]: I1203 16:24:37.780579 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Dec 03 16:24:37 crc kubenswrapper[4998]: I1203 16:24:37.818565 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Dec 03 16:24:37 crc kubenswrapper[4998]: I1203 16:24:37.823002 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-applier-0" Dec 03 16:24:37 crc kubenswrapper[4998]: I1203 16:24:37.881648 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-applier-0"] Dec 03 16:24:37 crc kubenswrapper[4998]: I1203 16:24:37.914967 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Dec 03 16:24:37 crc kubenswrapper[4998]: I1203 16:24:37.915038 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Dec 03 16:24:37 crc kubenswrapper[4998]: I1203 16:24:37.915124 4998 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 16:24:38 crc kubenswrapper[4998]: I1203 16:24:38.017322 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-67bd586fc-x2g2p" Dec 03 16:24:38 crc kubenswrapper[4998]: I1203 16:24:38.223018 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Dec 03 16:24:38 crc kubenswrapper[4998]: I1203 16:24:38.224552 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Dec 03 16:24:38 crc kubenswrapper[4998]: I1203 16:24:38.803091 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Dec 03 16:24:38 crc kubenswrapper[4998]: I1203 16:24:38.809119 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Dec 03 16:24:38 crc kubenswrapper[4998]: I1203 16:24:38.853089 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Dec 03 16:24:38 crc kubenswrapper[4998]: I1203 16:24:38.902717 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 16:24:39 crc kubenswrapper[4998]: I1203 16:24:39.809334 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-applier-0" podUID="5b94f565-0948-4e4f-9636-2833d43bb829" containerName="watcher-applier" containerID="cri-o://2c4f81d00299bd0d05b7f51e1c0c848a779a1b0b6f199befff2e205150316838" gracePeriod=30 Dec 03 16:24:40 crc kubenswrapper[4998]: I1203 16:24:40.819674 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-decision-engine-0" podUID="8e3693fc-394d-4cbb-8501-34bbc0483c37" containerName="watcher-decision-engine" containerID="cri-o://7de1267e2d008561febdc4f064e793bf1f979aae3c60266323b9c300d74fd5bd" gracePeriod=30 Dec 03 16:24:40 crc kubenswrapper[4998]: I1203 16:24:40.820064 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4faa24a-a43f-4d24-b20d-09e241d0cc96","Type":"ContainerStarted","Data":"a9100307a749acaf5cb3ca1c61dbd1c3014727d4fc73d649fe5f2633f22ad98e"} Dec 03 16:24:41 crc kubenswrapper[4998]: I1203 16:24:41.611203 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Dec 03 16:24:41 crc kubenswrapper[4998]: I1203 16:24:41.828811 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="41e16c37-8ac4-40c7-8873-fcb8364f31dc" containerName="watcher-api" containerID="cri-o://528c23fd97116c85c9f3479ca1d05476298c5e14f0628192a47e3177573979c4" gracePeriod=30 Dec 03 16:24:41 crc kubenswrapper[4998]: I1203 16:24:41.828741 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="41e16c37-8ac4-40c7-8873-fcb8364f31dc" containerName="watcher-api-log" containerID="cri-o://9035151ef65f166074aafe72a2fa23bdf67f54d4200dd95dc8aa8199cee7ca2b" gracePeriod=30 Dec 03 16:24:42 crc kubenswrapper[4998]: E1203 16:24:42.463368 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2c4f81d00299bd0d05b7f51e1c0c848a779a1b0b6f199befff2e205150316838" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 16:24:42 crc kubenswrapper[4998]: E1203 16:24:42.466056 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2c4f81d00299bd0d05b7f51e1c0c848a779a1b0b6f199befff2e205150316838" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 16:24:42 crc kubenswrapper[4998]: E1203 16:24:42.467731 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2c4f81d00299bd0d05b7f51e1c0c848a779a1b0b6f199befff2e205150316838" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 16:24:42 crc kubenswrapper[4998]: E1203 16:24:42.467791 4998 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/watcher-applier-0" podUID="5b94f565-0948-4e4f-9636-2833d43bb829" containerName="watcher-applier" Dec 03 16:24:42 crc kubenswrapper[4998]: I1203 16:24:42.915199 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="41e16c37-8ac4-40c7-8873-fcb8364f31dc" containerName="watcher-api-log" probeResult="failure" output="Get \"http://10.217.0.162:9322/\": dial tcp 10.217.0.162:9322: connect: connection refused" Dec 03 16:24:42 crc kubenswrapper[4998]: I1203 16:24:42.915350 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="41e16c37-8ac4-40c7-8873-fcb8364f31dc" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.162:9322/\": dial tcp 10.217.0.162:9322: connect: connection refused" Dec 03 16:24:43 crc kubenswrapper[4998]: E1203 16:24:43.678981 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.32:5001/podified-master-centos10/openstack-glance-api:watcher_latest\\\"\"" pod="openstack/glance-db-sync-2rqhm" podUID="de77cff6-7039-418c-8114-3c64ea4ce8e4" Dec 03 16:24:43 crc kubenswrapper[4998]: I1203 16:24:43.881927 4998 generic.go:334] "Generic (PLEG): container finished" podID="41e16c37-8ac4-40c7-8873-fcb8364f31dc" containerID="528c23fd97116c85c9f3479ca1d05476298c5e14f0628192a47e3177573979c4" exitCode=0 Dec 03 16:24:43 crc kubenswrapper[4998]: I1203 16:24:43.881958 4998 generic.go:334] "Generic (PLEG): container finished" podID="41e16c37-8ac4-40c7-8873-fcb8364f31dc" containerID="9035151ef65f166074aafe72a2fa23bdf67f54d4200dd95dc8aa8199cee7ca2b" exitCode=143 Dec 03 16:24:43 crc kubenswrapper[4998]: I1203 16:24:43.881984 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"41e16c37-8ac4-40c7-8873-fcb8364f31dc","Type":"ContainerDied","Data":"528c23fd97116c85c9f3479ca1d05476298c5e14f0628192a47e3177573979c4"} Dec 03 16:24:43 crc kubenswrapper[4998]: I1203 16:24:43.882013 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"41e16c37-8ac4-40c7-8873-fcb8364f31dc","Type":"ContainerDied","Data":"9035151ef65f166074aafe72a2fa23bdf67f54d4200dd95dc8aa8199cee7ca2b"} Dec 03 16:24:44 crc kubenswrapper[4998]: I1203 16:24:44.148246 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 03 16:24:44 crc kubenswrapper[4998]: I1203 16:24:44.238970 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hx79k\" (UniqueName: \"kubernetes.io/projected/41e16c37-8ac4-40c7-8873-fcb8364f31dc-kube-api-access-hx79k\") pod \"41e16c37-8ac4-40c7-8873-fcb8364f31dc\" (UID: \"41e16c37-8ac4-40c7-8873-fcb8364f31dc\") " Dec 03 16:24:44 crc kubenswrapper[4998]: I1203 16:24:44.239031 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41e16c37-8ac4-40c7-8873-fcb8364f31dc-combined-ca-bundle\") pod \"41e16c37-8ac4-40c7-8873-fcb8364f31dc\" (UID: \"41e16c37-8ac4-40c7-8873-fcb8364f31dc\") " Dec 03 16:24:44 crc kubenswrapper[4998]: I1203 16:24:44.239073 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41e16c37-8ac4-40c7-8873-fcb8364f31dc-config-data\") pod \"41e16c37-8ac4-40c7-8873-fcb8364f31dc\" (UID: \"41e16c37-8ac4-40c7-8873-fcb8364f31dc\") " Dec 03 16:24:44 crc kubenswrapper[4998]: I1203 16:24:44.239227 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41e16c37-8ac4-40c7-8873-fcb8364f31dc-logs\") pod \"41e16c37-8ac4-40c7-8873-fcb8364f31dc\" (UID: \"41e16c37-8ac4-40c7-8873-fcb8364f31dc\") " Dec 03 16:24:44 crc kubenswrapper[4998]: I1203 16:24:44.239303 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/41e16c37-8ac4-40c7-8873-fcb8364f31dc-custom-prometheus-ca\") pod \"41e16c37-8ac4-40c7-8873-fcb8364f31dc\" (UID: \"41e16c37-8ac4-40c7-8873-fcb8364f31dc\") " Dec 03 16:24:44 crc kubenswrapper[4998]: I1203 16:24:44.239855 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41e16c37-8ac4-40c7-8873-fcb8364f31dc-logs" (OuterVolumeSpecName: "logs") pod "41e16c37-8ac4-40c7-8873-fcb8364f31dc" (UID: "41e16c37-8ac4-40c7-8873-fcb8364f31dc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:24:44 crc kubenswrapper[4998]: I1203 16:24:44.280724 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41e16c37-8ac4-40c7-8873-fcb8364f31dc-kube-api-access-hx79k" (OuterVolumeSpecName: "kube-api-access-hx79k") pod "41e16c37-8ac4-40c7-8873-fcb8364f31dc" (UID: "41e16c37-8ac4-40c7-8873-fcb8364f31dc"). InnerVolumeSpecName "kube-api-access-hx79k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:24:44 crc kubenswrapper[4998]: I1203 16:24:44.316660 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41e16c37-8ac4-40c7-8873-fcb8364f31dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41e16c37-8ac4-40c7-8873-fcb8364f31dc" (UID: "41e16c37-8ac4-40c7-8873-fcb8364f31dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:24:44 crc kubenswrapper[4998]: I1203 16:24:44.316697 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41e16c37-8ac4-40c7-8873-fcb8364f31dc-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "41e16c37-8ac4-40c7-8873-fcb8364f31dc" (UID: "41e16c37-8ac4-40c7-8873-fcb8364f31dc"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:24:44 crc kubenswrapper[4998]: I1203 16:24:44.340939 4998 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41e16c37-8ac4-40c7-8873-fcb8364f31dc-logs\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:44 crc kubenswrapper[4998]: I1203 16:24:44.340971 4998 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/41e16c37-8ac4-40c7-8873-fcb8364f31dc-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:44 crc kubenswrapper[4998]: I1203 16:24:44.340985 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hx79k\" (UniqueName: \"kubernetes.io/projected/41e16c37-8ac4-40c7-8873-fcb8364f31dc-kube-api-access-hx79k\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:44 crc kubenswrapper[4998]: I1203 16:24:44.340996 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41e16c37-8ac4-40c7-8873-fcb8364f31dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:44 crc kubenswrapper[4998]: I1203 16:24:44.362964 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41e16c37-8ac4-40c7-8873-fcb8364f31dc-config-data" (OuterVolumeSpecName: "config-data") pod "41e16c37-8ac4-40c7-8873-fcb8364f31dc" (UID: "41e16c37-8ac4-40c7-8873-fcb8364f31dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:24:44 crc kubenswrapper[4998]: I1203 16:24:44.442453 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41e16c37-8ac4-40c7-8873-fcb8364f31dc-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:44 crc kubenswrapper[4998]: I1203 16:24:44.897502 4998 generic.go:334] "Generic (PLEG): container finished" podID="751f7a77-ec1d-4529-8762-a48fe772c25b" containerID="3e0b15332e4f017b612e36e98b7d4c7901623e0813d95832b81628ee8a6bbaeb" exitCode=0 Dec 03 16:24:44 crc kubenswrapper[4998]: I1203 16:24:44.897557 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bf2md" event={"ID":"751f7a77-ec1d-4529-8762-a48fe772c25b","Type":"ContainerDied","Data":"3e0b15332e4f017b612e36e98b7d4c7901623e0813d95832b81628ee8a6bbaeb"} Dec 03 16:24:44 crc kubenswrapper[4998]: I1203 16:24:44.902047 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"41e16c37-8ac4-40c7-8873-fcb8364f31dc","Type":"ContainerDied","Data":"e134225c8bcc72bb0f67575ff29b1c2c42310ace7087102fd3361dc42d90d7b6"} Dec 03 16:24:44 crc kubenswrapper[4998]: I1203 16:24:44.902078 4998 scope.go:117] "RemoveContainer" containerID="528c23fd97116c85c9f3479ca1d05476298c5e14f0628192a47e3177573979c4" Dec 03 16:24:44 crc kubenswrapper[4998]: I1203 16:24:44.902266 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 03 16:24:44 crc kubenswrapper[4998]: I1203 16:24:44.967093 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Dec 03 16:24:44 crc kubenswrapper[4998]: I1203 16:24:44.970459 4998 scope.go:117] "RemoveContainer" containerID="9035151ef65f166074aafe72a2fa23bdf67f54d4200dd95dc8aa8199cee7ca2b" Dec 03 16:24:44 crc kubenswrapper[4998]: I1203 16:24:44.995026 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-api-0"] Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.002023 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Dec 03 16:24:45 crc kubenswrapper[4998]: E1203 16:24:45.002639 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5c15d74-1721-40f9-bb22-162a95fd98c1" containerName="init" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.002659 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5c15d74-1721-40f9-bb22-162a95fd98c1" containerName="init" Dec 03 16:24:45 crc kubenswrapper[4998]: E1203 16:24:45.002680 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41e16c37-8ac4-40c7-8873-fcb8364f31dc" containerName="watcher-api-log" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.002687 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="41e16c37-8ac4-40c7-8873-fcb8364f31dc" containerName="watcher-api-log" Dec 03 16:24:45 crc kubenswrapper[4998]: E1203 16:24:45.002809 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41e16c37-8ac4-40c7-8873-fcb8364f31dc" containerName="watcher-api" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.002820 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="41e16c37-8ac4-40c7-8873-fcb8364f31dc" containerName="watcher-api" Dec 03 16:24:45 crc kubenswrapper[4998]: E1203 16:24:45.002846 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5c15d74-1721-40f9-bb22-162a95fd98c1" containerName="dnsmasq-dns" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.002853 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5c15d74-1721-40f9-bb22-162a95fd98c1" containerName="dnsmasq-dns" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.003084 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="41e16c37-8ac4-40c7-8873-fcb8364f31dc" containerName="watcher-api-log" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.003097 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5c15d74-1721-40f9-bb22-162a95fd98c1" containerName="dnsmasq-dns" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.003113 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="41e16c37-8ac4-40c7-8873-fcb8364f31dc" containerName="watcher-api" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.004245 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.008924 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-internal-svc" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.010021 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-public-svc" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.013766 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.017272 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.054062 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baf35f82-8685-4b68-b525-511a2a75aa27-config-data\") pod \"watcher-api-0\" (UID: \"baf35f82-8685-4b68-b525-511a2a75aa27\") " pod="openstack/watcher-api-0" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.054347 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pg86\" (UniqueName: \"kubernetes.io/projected/baf35f82-8685-4b68-b525-511a2a75aa27-kube-api-access-8pg86\") pod \"watcher-api-0\" (UID: \"baf35f82-8685-4b68-b525-511a2a75aa27\") " pod="openstack/watcher-api-0" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.054393 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/baf35f82-8685-4b68-b525-511a2a75aa27-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"baf35f82-8685-4b68-b525-511a2a75aa27\") " pod="openstack/watcher-api-0" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.054445 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/baf35f82-8685-4b68-b525-511a2a75aa27-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"baf35f82-8685-4b68-b525-511a2a75aa27\") " pod="openstack/watcher-api-0" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.054479 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baf35f82-8685-4b68-b525-511a2a75aa27-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"baf35f82-8685-4b68-b525-511a2a75aa27\") " pod="openstack/watcher-api-0" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.054513 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/baf35f82-8685-4b68-b525-511a2a75aa27-public-tls-certs\") pod \"watcher-api-0\" (UID: \"baf35f82-8685-4b68-b525-511a2a75aa27\") " pod="openstack/watcher-api-0" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.054556 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/baf35f82-8685-4b68-b525-511a2a75aa27-logs\") pod \"watcher-api-0\" (UID: \"baf35f82-8685-4b68-b525-511a2a75aa27\") " pod="openstack/watcher-api-0" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.156398 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baf35f82-8685-4b68-b525-511a2a75aa27-config-data\") pod \"watcher-api-0\" (UID: \"baf35f82-8685-4b68-b525-511a2a75aa27\") " pod="openstack/watcher-api-0" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.156434 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pg86\" (UniqueName: \"kubernetes.io/projected/baf35f82-8685-4b68-b525-511a2a75aa27-kube-api-access-8pg86\") pod \"watcher-api-0\" (UID: \"baf35f82-8685-4b68-b525-511a2a75aa27\") " pod="openstack/watcher-api-0" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.156468 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/baf35f82-8685-4b68-b525-511a2a75aa27-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"baf35f82-8685-4b68-b525-511a2a75aa27\") " pod="openstack/watcher-api-0" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.156511 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/baf35f82-8685-4b68-b525-511a2a75aa27-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"baf35f82-8685-4b68-b525-511a2a75aa27\") " pod="openstack/watcher-api-0" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.156541 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baf35f82-8685-4b68-b525-511a2a75aa27-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"baf35f82-8685-4b68-b525-511a2a75aa27\") " pod="openstack/watcher-api-0" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.156573 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/baf35f82-8685-4b68-b525-511a2a75aa27-public-tls-certs\") pod \"watcher-api-0\" (UID: \"baf35f82-8685-4b68-b525-511a2a75aa27\") " pod="openstack/watcher-api-0" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.156612 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/baf35f82-8685-4b68-b525-511a2a75aa27-logs\") pod \"watcher-api-0\" (UID: \"baf35f82-8685-4b68-b525-511a2a75aa27\") " pod="openstack/watcher-api-0" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.157183 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/baf35f82-8685-4b68-b525-511a2a75aa27-logs\") pod \"watcher-api-0\" (UID: \"baf35f82-8685-4b68-b525-511a2a75aa27\") " pod="openstack/watcher-api-0" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.161130 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baf35f82-8685-4b68-b525-511a2a75aa27-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"baf35f82-8685-4b68-b525-511a2a75aa27\") " pod="openstack/watcher-api-0" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.162482 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baf35f82-8685-4b68-b525-511a2a75aa27-config-data\") pod \"watcher-api-0\" (UID: \"baf35f82-8685-4b68-b525-511a2a75aa27\") " pod="openstack/watcher-api-0" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.166178 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/baf35f82-8685-4b68-b525-511a2a75aa27-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"baf35f82-8685-4b68-b525-511a2a75aa27\") " pod="openstack/watcher-api-0" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.166514 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/baf35f82-8685-4b68-b525-511a2a75aa27-public-tls-certs\") pod \"watcher-api-0\" (UID: \"baf35f82-8685-4b68-b525-511a2a75aa27\") " pod="openstack/watcher-api-0" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.167427 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/baf35f82-8685-4b68-b525-511a2a75aa27-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"baf35f82-8685-4b68-b525-511a2a75aa27\") " pod="openstack/watcher-api-0" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.176374 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pg86\" (UniqueName: \"kubernetes.io/projected/baf35f82-8685-4b68-b525-511a2a75aa27-kube-api-access-8pg86\") pod \"watcher-api-0\" (UID: \"baf35f82-8685-4b68-b525-511a2a75aa27\") " pod="openstack/watcher-api-0" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.348908 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.688830 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41e16c37-8ac4-40c7-8873-fcb8364f31dc" path="/var/lib/kubelet/pods/41e16c37-8ac4-40c7-8873-fcb8364f31dc/volumes" Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.848362 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Dec 03 16:24:45 crc kubenswrapper[4998]: I1203 16:24:45.915456 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"baf35f82-8685-4b68-b525-511a2a75aa27","Type":"ContainerStarted","Data":"8d071b109b8dc5a2aece5e3812fa7dba4079f56df03f1a22612953e860327310"} Dec 03 16:24:46 crc kubenswrapper[4998]: I1203 16:24:46.370591 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bf2md" Dec 03 16:24:46 crc kubenswrapper[4998]: I1203 16:24:46.486627 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-credential-keys\") pod \"751f7a77-ec1d-4529-8762-a48fe772c25b\" (UID: \"751f7a77-ec1d-4529-8762-a48fe772c25b\") " Dec 03 16:24:46 crc kubenswrapper[4998]: I1203 16:24:46.486734 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-fernet-keys\") pod \"751f7a77-ec1d-4529-8762-a48fe772c25b\" (UID: \"751f7a77-ec1d-4529-8762-a48fe772c25b\") " Dec 03 16:24:46 crc kubenswrapper[4998]: I1203 16:24:46.486863 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-config-data\") pod \"751f7a77-ec1d-4529-8762-a48fe772c25b\" (UID: \"751f7a77-ec1d-4529-8762-a48fe772c25b\") " Dec 03 16:24:46 crc kubenswrapper[4998]: I1203 16:24:46.486962 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-scripts\") pod \"751f7a77-ec1d-4529-8762-a48fe772c25b\" (UID: \"751f7a77-ec1d-4529-8762-a48fe772c25b\") " Dec 03 16:24:46 crc kubenswrapper[4998]: I1203 16:24:46.487069 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-combined-ca-bundle\") pod \"751f7a77-ec1d-4529-8762-a48fe772c25b\" (UID: \"751f7a77-ec1d-4529-8762-a48fe772c25b\") " Dec 03 16:24:46 crc kubenswrapper[4998]: I1203 16:24:46.487138 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mq4jd\" (UniqueName: \"kubernetes.io/projected/751f7a77-ec1d-4529-8762-a48fe772c25b-kube-api-access-mq4jd\") pod \"751f7a77-ec1d-4529-8762-a48fe772c25b\" (UID: \"751f7a77-ec1d-4529-8762-a48fe772c25b\") " Dec 03 16:24:46 crc kubenswrapper[4998]: I1203 16:24:46.494914 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "751f7a77-ec1d-4529-8762-a48fe772c25b" (UID: "751f7a77-ec1d-4529-8762-a48fe772c25b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:24:46 crc kubenswrapper[4998]: I1203 16:24:46.494970 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-scripts" (OuterVolumeSpecName: "scripts") pod "751f7a77-ec1d-4529-8762-a48fe772c25b" (UID: "751f7a77-ec1d-4529-8762-a48fe772c25b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:24:46 crc kubenswrapper[4998]: I1203 16:24:46.496963 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "751f7a77-ec1d-4529-8762-a48fe772c25b" (UID: "751f7a77-ec1d-4529-8762-a48fe772c25b"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:24:46 crc kubenswrapper[4998]: I1203 16:24:46.496990 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/751f7a77-ec1d-4529-8762-a48fe772c25b-kube-api-access-mq4jd" (OuterVolumeSpecName: "kube-api-access-mq4jd") pod "751f7a77-ec1d-4529-8762-a48fe772c25b" (UID: "751f7a77-ec1d-4529-8762-a48fe772c25b"). InnerVolumeSpecName "kube-api-access-mq4jd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:24:46 crc kubenswrapper[4998]: I1203 16:24:46.533782 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-config-data" (OuterVolumeSpecName: "config-data") pod "751f7a77-ec1d-4529-8762-a48fe772c25b" (UID: "751f7a77-ec1d-4529-8762-a48fe772c25b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:24:46 crc kubenswrapper[4998]: I1203 16:24:46.548904 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "751f7a77-ec1d-4529-8762-a48fe772c25b" (UID: "751f7a77-ec1d-4529-8762-a48fe772c25b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:24:46 crc kubenswrapper[4998]: I1203 16:24:46.589942 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:46 crc kubenswrapper[4998]: I1203 16:24:46.589974 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mq4jd\" (UniqueName: \"kubernetes.io/projected/751f7a77-ec1d-4529-8762-a48fe772c25b-kube-api-access-mq4jd\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:46 crc kubenswrapper[4998]: I1203 16:24:46.589987 4998 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:46 crc kubenswrapper[4998]: I1203 16:24:46.589996 4998 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:46 crc kubenswrapper[4998]: I1203 16:24:46.590006 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:46 crc kubenswrapper[4998]: I1203 16:24:46.590015 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/751f7a77-ec1d-4529-8762-a48fe772c25b-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:46 crc kubenswrapper[4998]: I1203 16:24:46.599966 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7565d78f8d-jm9qt" podUID="d7600639-7175-4c0a-9524-40adb505d3ea" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.159:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.159:8443: connect: connection refused" Dec 03 16:24:46 crc kubenswrapper[4998]: I1203 16:24:46.671700 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c5f7f9446-f97x8" podUID="24f1f198-c11a-4b5d-bcdd-d359ed88d97f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.160:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.160:8443: connect: connection refused" Dec 03 16:24:46 crc kubenswrapper[4998]: I1203 16:24:46.937723 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"baf35f82-8685-4b68-b525-511a2a75aa27","Type":"ContainerStarted","Data":"49187cf9396a0b5629766eae8b3bd36063d47d6c8a65b2e7c62bb2ee456c16af"} Dec 03 16:24:46 crc kubenswrapper[4998]: I1203 16:24:46.944233 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-bf2md" event={"ID":"751f7a77-ec1d-4529-8762-a48fe772c25b","Type":"ContainerDied","Data":"b0b836f41037c77bb86a6126a532b972f0cc30ed09612bcdc243c389f93c792f"} Dec 03 16:24:46 crc kubenswrapper[4998]: I1203 16:24:46.944285 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0b836f41037c77bb86a6126a532b972f0cc30ed09612bcdc243c389f93c792f" Dec 03 16:24:46 crc kubenswrapper[4998]: I1203 16:24:46.944292 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-bf2md" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.021775 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6764c5c65f-jtlz7"] Dec 03 16:24:47 crc kubenswrapper[4998]: E1203 16:24:47.031229 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="751f7a77-ec1d-4529-8762-a48fe772c25b" containerName="keystone-bootstrap" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.031382 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="751f7a77-ec1d-4529-8762-a48fe772c25b" containerName="keystone-bootstrap" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.036064 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="751f7a77-ec1d-4529-8762-a48fe772c25b" containerName="keystone-bootstrap" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.040731 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.043541 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.043723 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.051392 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6764c5c65f-jtlz7"] Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.052240 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.052424 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.052534 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.052559 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ggfph" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.098562 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d245e0d1-7808-4614-870e-64947ae9fb77-credential-keys\") pod \"keystone-6764c5c65f-jtlz7\" (UID: \"d245e0d1-7808-4614-870e-64947ae9fb77\") " pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.098722 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d245e0d1-7808-4614-870e-64947ae9fb77-public-tls-certs\") pod \"keystone-6764c5c65f-jtlz7\" (UID: \"d245e0d1-7808-4614-870e-64947ae9fb77\") " pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.098770 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d245e0d1-7808-4614-870e-64947ae9fb77-config-data\") pod \"keystone-6764c5c65f-jtlz7\" (UID: \"d245e0d1-7808-4614-870e-64947ae9fb77\") " pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.098799 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g6nd\" (UniqueName: \"kubernetes.io/projected/d245e0d1-7808-4614-870e-64947ae9fb77-kube-api-access-9g6nd\") pod \"keystone-6764c5c65f-jtlz7\" (UID: \"d245e0d1-7808-4614-870e-64947ae9fb77\") " pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.098833 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d245e0d1-7808-4614-870e-64947ae9fb77-combined-ca-bundle\") pod \"keystone-6764c5c65f-jtlz7\" (UID: \"d245e0d1-7808-4614-870e-64947ae9fb77\") " pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.098884 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d245e0d1-7808-4614-870e-64947ae9fb77-fernet-keys\") pod \"keystone-6764c5c65f-jtlz7\" (UID: \"d245e0d1-7808-4614-870e-64947ae9fb77\") " pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.098939 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d245e0d1-7808-4614-870e-64947ae9fb77-internal-tls-certs\") pod \"keystone-6764c5c65f-jtlz7\" (UID: \"d245e0d1-7808-4614-870e-64947ae9fb77\") " pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.098958 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d245e0d1-7808-4614-870e-64947ae9fb77-scripts\") pod \"keystone-6764c5c65f-jtlz7\" (UID: \"d245e0d1-7808-4614-870e-64947ae9fb77\") " pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.200249 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d245e0d1-7808-4614-870e-64947ae9fb77-combined-ca-bundle\") pod \"keystone-6764c5c65f-jtlz7\" (UID: \"d245e0d1-7808-4614-870e-64947ae9fb77\") " pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.200348 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d245e0d1-7808-4614-870e-64947ae9fb77-fernet-keys\") pod \"keystone-6764c5c65f-jtlz7\" (UID: \"d245e0d1-7808-4614-870e-64947ae9fb77\") " pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.200410 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d245e0d1-7808-4614-870e-64947ae9fb77-internal-tls-certs\") pod \"keystone-6764c5c65f-jtlz7\" (UID: \"d245e0d1-7808-4614-870e-64947ae9fb77\") " pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.200429 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d245e0d1-7808-4614-870e-64947ae9fb77-scripts\") pod \"keystone-6764c5c65f-jtlz7\" (UID: \"d245e0d1-7808-4614-870e-64947ae9fb77\") " pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.200467 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d245e0d1-7808-4614-870e-64947ae9fb77-credential-keys\") pod \"keystone-6764c5c65f-jtlz7\" (UID: \"d245e0d1-7808-4614-870e-64947ae9fb77\") " pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.200483 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d245e0d1-7808-4614-870e-64947ae9fb77-public-tls-certs\") pod \"keystone-6764c5c65f-jtlz7\" (UID: \"d245e0d1-7808-4614-870e-64947ae9fb77\") " pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.200510 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d245e0d1-7808-4614-870e-64947ae9fb77-config-data\") pod \"keystone-6764c5c65f-jtlz7\" (UID: \"d245e0d1-7808-4614-870e-64947ae9fb77\") " pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.200530 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g6nd\" (UniqueName: \"kubernetes.io/projected/d245e0d1-7808-4614-870e-64947ae9fb77-kube-api-access-9g6nd\") pod \"keystone-6764c5c65f-jtlz7\" (UID: \"d245e0d1-7808-4614-870e-64947ae9fb77\") " pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.210771 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d245e0d1-7808-4614-870e-64947ae9fb77-internal-tls-certs\") pod \"keystone-6764c5c65f-jtlz7\" (UID: \"d245e0d1-7808-4614-870e-64947ae9fb77\") " pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.212674 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d245e0d1-7808-4614-870e-64947ae9fb77-config-data\") pod \"keystone-6764c5c65f-jtlz7\" (UID: \"d245e0d1-7808-4614-870e-64947ae9fb77\") " pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.214002 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d245e0d1-7808-4614-870e-64947ae9fb77-scripts\") pod \"keystone-6764c5c65f-jtlz7\" (UID: \"d245e0d1-7808-4614-870e-64947ae9fb77\") " pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.214314 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d245e0d1-7808-4614-870e-64947ae9fb77-credential-keys\") pod \"keystone-6764c5c65f-jtlz7\" (UID: \"d245e0d1-7808-4614-870e-64947ae9fb77\") " pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.214537 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d245e0d1-7808-4614-870e-64947ae9fb77-fernet-keys\") pod \"keystone-6764c5c65f-jtlz7\" (UID: \"d245e0d1-7808-4614-870e-64947ae9fb77\") " pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.217149 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d245e0d1-7808-4614-870e-64947ae9fb77-combined-ca-bundle\") pod \"keystone-6764c5c65f-jtlz7\" (UID: \"d245e0d1-7808-4614-870e-64947ae9fb77\") " pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.217364 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d245e0d1-7808-4614-870e-64947ae9fb77-public-tls-certs\") pod \"keystone-6764c5c65f-jtlz7\" (UID: \"d245e0d1-7808-4614-870e-64947ae9fb77\") " pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.223854 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g6nd\" (UniqueName: \"kubernetes.io/projected/d245e0d1-7808-4614-870e-64947ae9fb77-kube-api-access-9g6nd\") pod \"keystone-6764c5c65f-jtlz7\" (UID: \"d245e0d1-7808-4614-870e-64947ae9fb77\") " pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.418052 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:47 crc kubenswrapper[4998]: E1203 16:24:47.463671 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2c4f81d00299bd0d05b7f51e1c0c848a779a1b0b6f199befff2e205150316838" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 16:24:47 crc kubenswrapper[4998]: E1203 16:24:47.467833 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2c4f81d00299bd0d05b7f51e1c0c848a779a1b0b6f199befff2e205150316838" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 16:24:47 crc kubenswrapper[4998]: E1203 16:24:47.475111 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2c4f81d00299bd0d05b7f51e1c0c848a779a1b0b6f199befff2e205150316838" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 16:24:47 crc kubenswrapper[4998]: E1203 16:24:47.475209 4998 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/watcher-applier-0" podUID="5b94f565-0948-4e4f-9636-2833d43bb829" containerName="watcher-applier" Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.922723 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6764c5c65f-jtlz7"] Dec 03 16:24:47 crc kubenswrapper[4998]: I1203 16:24:47.961918 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6764c5c65f-jtlz7" event={"ID":"d245e0d1-7808-4614-870e-64947ae9fb77","Type":"ContainerStarted","Data":"a2495d6d18a4909d62297e1b44f26ffd21a1a78398947a2f7503ed46089eb79c"} Dec 03 16:24:48 crc kubenswrapper[4998]: I1203 16:24:48.973269 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6764c5c65f-jtlz7" event={"ID":"d245e0d1-7808-4614-870e-64947ae9fb77","Type":"ContainerStarted","Data":"5c23d056060e2220760642debe686a698fab078d9df2d54a3e8bf14b6598dcc1"} Dec 03 16:24:48 crc kubenswrapper[4998]: I1203 16:24:48.975952 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"baf35f82-8685-4b68-b525-511a2a75aa27","Type":"ContainerStarted","Data":"ba48bd48d2352e5ed9a47f07e87c9eb63ca8385d95fcf85e27bd481b26606746"} Dec 03 16:24:49 crc kubenswrapper[4998]: I1203 16:24:49.983278 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Dec 03 16:24:49 crc kubenswrapper[4998]: I1203 16:24:49.983626 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:24:50 crc kubenswrapper[4998]: I1203 16:24:50.015266 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6764c5c65f-jtlz7" podStartSLOduration=4.015237852 podStartE2EDuration="4.015237852s" podCreationTimestamp="2025-12-03 16:24:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:24:50.01024289 +0000 UTC m=+1268.621943133" watchObservedRunningTime="2025-12-03 16:24:50.015237852 +0000 UTC m=+1268.626938075" Dec 03 16:24:50 crc kubenswrapper[4998]: I1203 16:24:50.349158 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Dec 03 16:24:52 crc kubenswrapper[4998]: I1203 16:24:52.004631 4998 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 16:24:52 crc kubenswrapper[4998]: I1203 16:24:52.243948 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Dec 03 16:24:52 crc kubenswrapper[4998]: I1203 16:24:52.272960 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=8.272939251 podStartE2EDuration="8.272939251s" podCreationTimestamp="2025-12-03 16:24:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:24:50.042045644 +0000 UTC m=+1268.653745867" watchObservedRunningTime="2025-12-03 16:24:52.272939251 +0000 UTC m=+1270.884639484" Dec 03 16:24:52 crc kubenswrapper[4998]: E1203 16:24:52.463698 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2c4f81d00299bd0d05b7f51e1c0c848a779a1b0b6f199befff2e205150316838" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 16:24:52 crc kubenswrapper[4998]: E1203 16:24:52.465415 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2c4f81d00299bd0d05b7f51e1c0c848a779a1b0b6f199befff2e205150316838" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 16:24:52 crc kubenswrapper[4998]: E1203 16:24:52.467167 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2c4f81d00299bd0d05b7f51e1c0c848a779a1b0b6f199befff2e205150316838" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 16:24:52 crc kubenswrapper[4998]: E1203 16:24:52.467214 4998 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/watcher-applier-0" podUID="5b94f565-0948-4e4f-9636-2833d43bb829" containerName="watcher-applier" Dec 03 16:24:55 crc kubenswrapper[4998]: I1203 16:24:55.349094 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Dec 03 16:24:55 crc kubenswrapper[4998]: I1203 16:24:55.355375 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Dec 03 16:24:56 crc kubenswrapper[4998]: I1203 16:24:56.097495 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Dec 03 16:24:57 crc kubenswrapper[4998]: I1203 16:24:57.110851 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:24:57 crc kubenswrapper[4998]: I1203 16:24:57.110914 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:24:57 crc kubenswrapper[4998]: E1203 16:24:57.466012 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2c4f81d00299bd0d05b7f51e1c0c848a779a1b0b6f199befff2e205150316838" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 16:24:57 crc kubenswrapper[4998]: E1203 16:24:57.468572 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2c4f81d00299bd0d05b7f51e1c0c848a779a1b0b6f199befff2e205150316838" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 16:24:57 crc kubenswrapper[4998]: E1203 16:24:57.470597 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2c4f81d00299bd0d05b7f51e1c0c848a779a1b0b6f199befff2e205150316838" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 16:24:57 crc kubenswrapper[4998]: E1203 16:24:57.470704 4998 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/watcher-applier-0" podUID="5b94f565-0948-4e4f-9636-2833d43bb829" containerName="watcher-applier" Dec 03 16:24:57 crc kubenswrapper[4998]: E1203 16:24:57.780556 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7de1267e2d008561febdc4f064e793bf1f979aae3c60266323b9c300d74fd5bd is running failed: container process not found" containerID="7de1267e2d008561febdc4f064e793bf1f979aae3c60266323b9c300d74fd5bd" cmd=["/usr/bin/pgrep","-f","-r","DRST","watcher-decision-engine"] Dec 03 16:24:57 crc kubenswrapper[4998]: E1203 16:24:57.780976 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7de1267e2d008561febdc4f064e793bf1f979aae3c60266323b9c300d74fd5bd is running failed: container process not found" containerID="7de1267e2d008561febdc4f064e793bf1f979aae3c60266323b9c300d74fd5bd" cmd=["/usr/bin/pgrep","-f","-r","DRST","watcher-decision-engine"] Dec 03 16:24:57 crc kubenswrapper[4998]: E1203 16:24:57.781369 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7de1267e2d008561febdc4f064e793bf1f979aae3c60266323b9c300d74fd5bd is running failed: container process not found" containerID="7de1267e2d008561febdc4f064e793bf1f979aae3c60266323b9c300d74fd5bd" cmd=["/usr/bin/pgrep","-f","-r","DRST","watcher-decision-engine"] Dec 03 16:24:57 crc kubenswrapper[4998]: E1203 16:24:57.781427 4998 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7de1267e2d008561febdc4f064e793bf1f979aae3c60266323b9c300d74fd5bd is running failed: container process not found" probeType="Readiness" pod="openstack/watcher-decision-engine-0" podUID="8e3693fc-394d-4cbb-8501-34bbc0483c37" containerName="watcher-decision-engine" Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.062853 4998 generic.go:334] "Generic (PLEG): container finished" podID="5cca43ea-2244-4f5c-8704-8d91634c0ad6" containerID="a2dafe752937450125658f1ce9546da3c4d740014f80b43ed2fba0e66e8931f8" exitCode=0 Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.063071 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qw92l" event={"ID":"5cca43ea-2244-4f5c-8704-8d91634c0ad6","Type":"ContainerDied","Data":"a2dafe752937450125658f1ce9546da3c4d740014f80b43ed2fba0e66e8931f8"} Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.066009 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-w8k5f" event={"ID":"5d2429d8-9baa-4532-9f8b-9ffb72a3a764","Type":"ContainerStarted","Data":"9c2779da866bb38d647d67f9337f9006e0141dc294f2b196b28cf1e700822fcf"} Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.070131 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4faa24a-a43f-4d24-b20d-09e241d0cc96","Type":"ContainerStarted","Data":"6f0d6533d024330f91e4fcc8a4ad9f21d7ed7fb31327d7ca16b483a8f88d0308"} Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.073290 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-2rqhm" event={"ID":"de77cff6-7039-418c-8114-3c64ea4ce8e4","Type":"ContainerStarted","Data":"043af9beddc6495b4dd97ae2a362ea5bea94ef2a3d8e62d695b5f5765c933eda"} Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.076830 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8lz9j" event={"ID":"82aa9d23-457c-45a0-9729-e5a803738237","Type":"ContainerStarted","Data":"2298bd80e5bdaea4315dbb4b6a0effa05cbdcae8e440ead4f9287b4f7bd6ed75"} Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.081441 4998 generic.go:334] "Generic (PLEG): container finished" podID="8e3693fc-394d-4cbb-8501-34bbc0483c37" containerID="7de1267e2d008561febdc4f064e793bf1f979aae3c60266323b9c300d74fd5bd" exitCode=1 Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.081495 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"8e3693fc-394d-4cbb-8501-34bbc0483c37","Type":"ContainerDied","Data":"7de1267e2d008561febdc4f064e793bf1f979aae3c60266323b9c300d74fd5bd"} Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.153010 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-w8k5f" podStartSLOduration=3.52596129 podStartE2EDuration="1m1.152982126s" podCreationTimestamp="2025-12-03 16:23:57 +0000 UTC" firstStartedPulling="2025-12-03 16:23:59.04605523 +0000 UTC m=+1217.657755453" lastFinishedPulling="2025-12-03 16:24:56.673076056 +0000 UTC m=+1275.284776289" observedRunningTime="2025-12-03 16:24:58.101172646 +0000 UTC m=+1276.712872869" watchObservedRunningTime="2025-12-03 16:24:58.152982126 +0000 UTC m=+1276.764682349" Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.159770 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-2rqhm" podStartSLOduration=4.233413246 podStartE2EDuration="1m24.15972687s" podCreationTimestamp="2025-12-03 16:23:34 +0000 UTC" firstStartedPulling="2025-12-03 16:23:36.87615188 +0000 UTC m=+1195.487852123" lastFinishedPulling="2025-12-03 16:24:56.802465524 +0000 UTC m=+1275.414165747" observedRunningTime="2025-12-03 16:24:58.116849457 +0000 UTC m=+1276.728549690" watchObservedRunningTime="2025-12-03 16:24:58.15972687 +0000 UTC m=+1276.771427113" Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.176958 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-8lz9j" podStartSLOduration=3.455062693 podStartE2EDuration="1m1.176937009s" podCreationTimestamp="2025-12-03 16:23:57 +0000 UTC" firstStartedPulling="2025-12-03 16:23:58.951233691 +0000 UTC m=+1217.562933914" lastFinishedPulling="2025-12-03 16:24:56.673107987 +0000 UTC m=+1275.284808230" observedRunningTime="2025-12-03 16:24:58.139033737 +0000 UTC m=+1276.750733980" watchObservedRunningTime="2025-12-03 16:24:58.176937009 +0000 UTC m=+1276.788637242" Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.218872 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.354059 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e3693fc-394d-4cbb-8501-34bbc0483c37-logs\") pod \"8e3693fc-394d-4cbb-8501-34bbc0483c37\" (UID: \"8e3693fc-394d-4cbb-8501-34bbc0483c37\") " Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.354114 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqkgk\" (UniqueName: \"kubernetes.io/projected/8e3693fc-394d-4cbb-8501-34bbc0483c37-kube-api-access-xqkgk\") pod \"8e3693fc-394d-4cbb-8501-34bbc0483c37\" (UID: \"8e3693fc-394d-4cbb-8501-34bbc0483c37\") " Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.354192 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/8e3693fc-394d-4cbb-8501-34bbc0483c37-custom-prometheus-ca\") pod \"8e3693fc-394d-4cbb-8501-34bbc0483c37\" (UID: \"8e3693fc-394d-4cbb-8501-34bbc0483c37\") " Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.354247 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e3693fc-394d-4cbb-8501-34bbc0483c37-config-data\") pod \"8e3693fc-394d-4cbb-8501-34bbc0483c37\" (UID: \"8e3693fc-394d-4cbb-8501-34bbc0483c37\") " Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.354373 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e3693fc-394d-4cbb-8501-34bbc0483c37-combined-ca-bundle\") pod \"8e3693fc-394d-4cbb-8501-34bbc0483c37\" (UID: \"8e3693fc-394d-4cbb-8501-34bbc0483c37\") " Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.356743 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e3693fc-394d-4cbb-8501-34bbc0483c37-logs" (OuterVolumeSpecName: "logs") pod "8e3693fc-394d-4cbb-8501-34bbc0483c37" (UID: "8e3693fc-394d-4cbb-8501-34bbc0483c37"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.366053 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e3693fc-394d-4cbb-8501-34bbc0483c37-kube-api-access-xqkgk" (OuterVolumeSpecName: "kube-api-access-xqkgk") pod "8e3693fc-394d-4cbb-8501-34bbc0483c37" (UID: "8e3693fc-394d-4cbb-8501-34bbc0483c37"). InnerVolumeSpecName "kube-api-access-xqkgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.385238 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e3693fc-394d-4cbb-8501-34bbc0483c37-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8e3693fc-394d-4cbb-8501-34bbc0483c37" (UID: "8e3693fc-394d-4cbb-8501-34bbc0483c37"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.409716 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e3693fc-394d-4cbb-8501-34bbc0483c37-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "8e3693fc-394d-4cbb-8501-34bbc0483c37" (UID: "8e3693fc-394d-4cbb-8501-34bbc0483c37"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.427883 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e3693fc-394d-4cbb-8501-34bbc0483c37-config-data" (OuterVolumeSpecName: "config-data") pod "8e3693fc-394d-4cbb-8501-34bbc0483c37" (UID: "8e3693fc-394d-4cbb-8501-34bbc0483c37"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.456492 4998 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e3693fc-394d-4cbb-8501-34bbc0483c37-logs\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.456546 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqkgk\" (UniqueName: \"kubernetes.io/projected/8e3693fc-394d-4cbb-8501-34bbc0483c37-kube-api-access-xqkgk\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.456561 4998 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/8e3693fc-394d-4cbb-8501-34bbc0483c37-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.456577 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e3693fc-394d-4cbb-8501-34bbc0483c37-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.456591 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e3693fc-394d-4cbb-8501-34bbc0483c37-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.660509 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:24:58 crc kubenswrapper[4998]: I1203 16:24:58.670745 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.097491 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"8e3693fc-394d-4cbb-8501-34bbc0483c37","Type":"ContainerDied","Data":"0953a810f801d2d84d0ac7482bf35710bf063e49da93bce33dc04703dab5e740"} Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.097539 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.097575 4998 scope.go:117] "RemoveContainer" containerID="7de1267e2d008561febdc4f064e793bf1f979aae3c60266323b9c300d74fd5bd" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.138734 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.160044 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.175010 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 16:24:59 crc kubenswrapper[4998]: E1203 16:24:59.175363 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e3693fc-394d-4cbb-8501-34bbc0483c37" containerName="watcher-decision-engine" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.175375 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e3693fc-394d-4cbb-8501-34bbc0483c37" containerName="watcher-decision-engine" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.175567 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e3693fc-394d-4cbb-8501-34bbc0483c37" containerName="watcher-decision-engine" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.176130 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.181066 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.188622 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.276373 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcvvt\" (UniqueName: \"kubernetes.io/projected/a009aeee-c849-4c00-b41b-3cf5e36a75bb-kube-api-access-fcvvt\") pod \"watcher-decision-engine-0\" (UID: \"a009aeee-c849-4c00-b41b-3cf5e36a75bb\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.276415 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a009aeee-c849-4c00-b41b-3cf5e36a75bb-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"a009aeee-c849-4c00-b41b-3cf5e36a75bb\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.276470 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a009aeee-c849-4c00-b41b-3cf5e36a75bb-config-data\") pod \"watcher-decision-engine-0\" (UID: \"a009aeee-c849-4c00-b41b-3cf5e36a75bb\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.276987 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a009aeee-c849-4c00-b41b-3cf5e36a75bb-logs\") pod \"watcher-decision-engine-0\" (UID: \"a009aeee-c849-4c00-b41b-3cf5e36a75bb\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.277054 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/a009aeee-c849-4c00-b41b-3cf5e36a75bb-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"a009aeee-c849-4c00-b41b-3cf5e36a75bb\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.378771 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a009aeee-c849-4c00-b41b-3cf5e36a75bb-logs\") pod \"watcher-decision-engine-0\" (UID: \"a009aeee-c849-4c00-b41b-3cf5e36a75bb\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.378818 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/a009aeee-c849-4c00-b41b-3cf5e36a75bb-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"a009aeee-c849-4c00-b41b-3cf5e36a75bb\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.378863 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcvvt\" (UniqueName: \"kubernetes.io/projected/a009aeee-c849-4c00-b41b-3cf5e36a75bb-kube-api-access-fcvvt\") pod \"watcher-decision-engine-0\" (UID: \"a009aeee-c849-4c00-b41b-3cf5e36a75bb\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.378880 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a009aeee-c849-4c00-b41b-3cf5e36a75bb-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"a009aeee-c849-4c00-b41b-3cf5e36a75bb\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.378921 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a009aeee-c849-4c00-b41b-3cf5e36a75bb-config-data\") pod \"watcher-decision-engine-0\" (UID: \"a009aeee-c849-4c00-b41b-3cf5e36a75bb\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.387364 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a009aeee-c849-4c00-b41b-3cf5e36a75bb-logs\") pod \"watcher-decision-engine-0\" (UID: \"a009aeee-c849-4c00-b41b-3cf5e36a75bb\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.395321 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/a009aeee-c849-4c00-b41b-3cf5e36a75bb-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"a009aeee-c849-4c00-b41b-3cf5e36a75bb\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.401110 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcvvt\" (UniqueName: \"kubernetes.io/projected/a009aeee-c849-4c00-b41b-3cf5e36a75bb-kube-api-access-fcvvt\") pod \"watcher-decision-engine-0\" (UID: \"a009aeee-c849-4c00-b41b-3cf5e36a75bb\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.402946 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a009aeee-c849-4c00-b41b-3cf5e36a75bb-config-data\") pod \"watcher-decision-engine-0\" (UID: \"a009aeee-c849-4c00-b41b-3cf5e36a75bb\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.409906 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a009aeee-c849-4c00-b41b-3cf5e36a75bb-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"a009aeee-c849-4c00-b41b-3cf5e36a75bb\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.496059 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qw92l" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.514152 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.581432 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cca43ea-2244-4f5c-8704-8d91634c0ad6-combined-ca-bundle\") pod \"5cca43ea-2244-4f5c-8704-8d91634c0ad6\" (UID: \"5cca43ea-2244-4f5c-8704-8d91634c0ad6\") " Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.581536 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mk4d5\" (UniqueName: \"kubernetes.io/projected/5cca43ea-2244-4f5c-8704-8d91634c0ad6-kube-api-access-mk4d5\") pod \"5cca43ea-2244-4f5c-8704-8d91634c0ad6\" (UID: \"5cca43ea-2244-4f5c-8704-8d91634c0ad6\") " Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.581575 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cca43ea-2244-4f5c-8704-8d91634c0ad6-config-data\") pod \"5cca43ea-2244-4f5c-8704-8d91634c0ad6\" (UID: \"5cca43ea-2244-4f5c-8704-8d91634c0ad6\") " Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.581725 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cca43ea-2244-4f5c-8704-8d91634c0ad6-logs\") pod \"5cca43ea-2244-4f5c-8704-8d91634c0ad6\" (UID: \"5cca43ea-2244-4f5c-8704-8d91634c0ad6\") " Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.581826 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cca43ea-2244-4f5c-8704-8d91634c0ad6-scripts\") pod \"5cca43ea-2244-4f5c-8704-8d91634c0ad6\" (UID: \"5cca43ea-2244-4f5c-8704-8d91634c0ad6\") " Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.587126 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cca43ea-2244-4f5c-8704-8d91634c0ad6-logs" (OuterVolumeSpecName: "logs") pod "5cca43ea-2244-4f5c-8704-8d91634c0ad6" (UID: "5cca43ea-2244-4f5c-8704-8d91634c0ad6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.587466 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cca43ea-2244-4f5c-8704-8d91634c0ad6-kube-api-access-mk4d5" (OuterVolumeSpecName: "kube-api-access-mk4d5") pod "5cca43ea-2244-4f5c-8704-8d91634c0ad6" (UID: "5cca43ea-2244-4f5c-8704-8d91634c0ad6"). InnerVolumeSpecName "kube-api-access-mk4d5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.587578 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cca43ea-2244-4f5c-8704-8d91634c0ad6-scripts" (OuterVolumeSpecName: "scripts") pod "5cca43ea-2244-4f5c-8704-8d91634c0ad6" (UID: "5cca43ea-2244-4f5c-8704-8d91634c0ad6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.626929 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cca43ea-2244-4f5c-8704-8d91634c0ad6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5cca43ea-2244-4f5c-8704-8d91634c0ad6" (UID: "5cca43ea-2244-4f5c-8704-8d91634c0ad6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.634481 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cca43ea-2244-4f5c-8704-8d91634c0ad6-config-data" (OuterVolumeSpecName: "config-data") pod "5cca43ea-2244-4f5c-8704-8d91634c0ad6" (UID: "5cca43ea-2244-4f5c-8704-8d91634c0ad6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.684055 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mk4d5\" (UniqueName: \"kubernetes.io/projected/5cca43ea-2244-4f5c-8704-8d91634c0ad6-kube-api-access-mk4d5\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.684095 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cca43ea-2244-4f5c-8704-8d91634c0ad6-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.684109 4998 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cca43ea-2244-4f5c-8704-8d91634c0ad6-logs\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.684117 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cca43ea-2244-4f5c-8704-8d91634c0ad6-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.684127 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cca43ea-2244-4f5c-8704-8d91634c0ad6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.700258 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e3693fc-394d-4cbb-8501-34bbc0483c37" path="/var/lib/kubelet/pods/8e3693fc-394d-4cbb-8501-34bbc0483c37/volumes" Dec 03 16:24:59 crc kubenswrapper[4998]: I1203 16:24:59.952861 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 16:24:59 crc kubenswrapper[4998]: W1203 16:24:59.967943 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda009aeee_c849_4c00_b41b_3cf5e36a75bb.slice/crio-510d649767f5e1e8480ee19550b358b84232ec13dd0041d57c75b1d5574ef8b3 WatchSource:0}: Error finding container 510d649767f5e1e8480ee19550b358b84232ec13dd0041d57c75b1d5574ef8b3: Status 404 returned error can't find the container with id 510d649767f5e1e8480ee19550b358b84232ec13dd0041d57c75b1d5574ef8b3 Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.126985 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-qw92l" event={"ID":"5cca43ea-2244-4f5c-8704-8d91634c0ad6","Type":"ContainerDied","Data":"068164d44989316e3d60e0b6bc2fde5aaabb737063cf017e0f9900b141460798"} Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.127024 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="068164d44989316e3d60e0b6bc2fde5aaabb737063cf017e0f9900b141460798" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.127041 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-qw92l" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.128176 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"a009aeee-c849-4c00-b41b-3cf5e36a75bb","Type":"ContainerStarted","Data":"510d649767f5e1e8480ee19550b358b84232ec13dd0041d57c75b1d5574ef8b3"} Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.187916 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-57c877d99d-7qbcv"] Dec 03 16:25:00 crc kubenswrapper[4998]: E1203 16:25:00.188608 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cca43ea-2244-4f5c-8704-8d91634c0ad6" containerName="placement-db-sync" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.188627 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cca43ea-2244-4f5c-8704-8d91634c0ad6" containerName="placement-db-sync" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.188858 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cca43ea-2244-4f5c-8704-8d91634c0ad6" containerName="placement-db-sync" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.190144 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.193773 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.193997 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.194139 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-9r2gs" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.194722 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.203792 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.205708 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-57c877d99d-7qbcv"] Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.308884 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/050d52db-a22f-43de-bf07-ed0b6bed49db-config-data\") pod \"placement-57c877d99d-7qbcv\" (UID: \"050d52db-a22f-43de-bf07-ed0b6bed49db\") " pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.309204 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/050d52db-a22f-43de-bf07-ed0b6bed49db-public-tls-certs\") pod \"placement-57c877d99d-7qbcv\" (UID: \"050d52db-a22f-43de-bf07-ed0b6bed49db\") " pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.309229 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/050d52db-a22f-43de-bf07-ed0b6bed49db-combined-ca-bundle\") pod \"placement-57c877d99d-7qbcv\" (UID: \"050d52db-a22f-43de-bf07-ed0b6bed49db\") " pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.309296 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/050d52db-a22f-43de-bf07-ed0b6bed49db-logs\") pod \"placement-57c877d99d-7qbcv\" (UID: \"050d52db-a22f-43de-bf07-ed0b6bed49db\") " pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.309408 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/050d52db-a22f-43de-bf07-ed0b6bed49db-scripts\") pod \"placement-57c877d99d-7qbcv\" (UID: \"050d52db-a22f-43de-bf07-ed0b6bed49db\") " pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.309479 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/050d52db-a22f-43de-bf07-ed0b6bed49db-internal-tls-certs\") pod \"placement-57c877d99d-7qbcv\" (UID: \"050d52db-a22f-43de-bf07-ed0b6bed49db\") " pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.309601 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjmzx\" (UniqueName: \"kubernetes.io/projected/050d52db-a22f-43de-bf07-ed0b6bed49db-kube-api-access-kjmzx\") pod \"placement-57c877d99d-7qbcv\" (UID: \"050d52db-a22f-43de-bf07-ed0b6bed49db\") " pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.412073 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjmzx\" (UniqueName: \"kubernetes.io/projected/050d52db-a22f-43de-bf07-ed0b6bed49db-kube-api-access-kjmzx\") pod \"placement-57c877d99d-7qbcv\" (UID: \"050d52db-a22f-43de-bf07-ed0b6bed49db\") " pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.412196 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/050d52db-a22f-43de-bf07-ed0b6bed49db-config-data\") pod \"placement-57c877d99d-7qbcv\" (UID: \"050d52db-a22f-43de-bf07-ed0b6bed49db\") " pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.412254 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/050d52db-a22f-43de-bf07-ed0b6bed49db-public-tls-certs\") pod \"placement-57c877d99d-7qbcv\" (UID: \"050d52db-a22f-43de-bf07-ed0b6bed49db\") " pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.412276 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/050d52db-a22f-43de-bf07-ed0b6bed49db-combined-ca-bundle\") pod \"placement-57c877d99d-7qbcv\" (UID: \"050d52db-a22f-43de-bf07-ed0b6bed49db\") " pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.412327 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/050d52db-a22f-43de-bf07-ed0b6bed49db-logs\") pod \"placement-57c877d99d-7qbcv\" (UID: \"050d52db-a22f-43de-bf07-ed0b6bed49db\") " pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.412362 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/050d52db-a22f-43de-bf07-ed0b6bed49db-scripts\") pod \"placement-57c877d99d-7qbcv\" (UID: \"050d52db-a22f-43de-bf07-ed0b6bed49db\") " pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.412654 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/050d52db-a22f-43de-bf07-ed0b6bed49db-internal-tls-certs\") pod \"placement-57c877d99d-7qbcv\" (UID: \"050d52db-a22f-43de-bf07-ed0b6bed49db\") " pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.413209 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/050d52db-a22f-43de-bf07-ed0b6bed49db-logs\") pod \"placement-57c877d99d-7qbcv\" (UID: \"050d52db-a22f-43de-bf07-ed0b6bed49db\") " pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.417522 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/050d52db-a22f-43de-bf07-ed0b6bed49db-scripts\") pod \"placement-57c877d99d-7qbcv\" (UID: \"050d52db-a22f-43de-bf07-ed0b6bed49db\") " pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.417965 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/050d52db-a22f-43de-bf07-ed0b6bed49db-public-tls-certs\") pod \"placement-57c877d99d-7qbcv\" (UID: \"050d52db-a22f-43de-bf07-ed0b6bed49db\") " pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.418130 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/050d52db-a22f-43de-bf07-ed0b6bed49db-combined-ca-bundle\") pod \"placement-57c877d99d-7qbcv\" (UID: \"050d52db-a22f-43de-bf07-ed0b6bed49db\") " pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.418570 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/050d52db-a22f-43de-bf07-ed0b6bed49db-config-data\") pod \"placement-57c877d99d-7qbcv\" (UID: \"050d52db-a22f-43de-bf07-ed0b6bed49db\") " pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.427884 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/050d52db-a22f-43de-bf07-ed0b6bed49db-internal-tls-certs\") pod \"placement-57c877d99d-7qbcv\" (UID: \"050d52db-a22f-43de-bf07-ed0b6bed49db\") " pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.431448 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjmzx\" (UniqueName: \"kubernetes.io/projected/050d52db-a22f-43de-bf07-ed0b6bed49db-kube-api-access-kjmzx\") pod \"placement-57c877d99d-7qbcv\" (UID: \"050d52db-a22f-43de-bf07-ed0b6bed49db\") " pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.513461 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:00 crc kubenswrapper[4998]: I1203 16:25:00.833350 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-57c877d99d-7qbcv"] Dec 03 16:25:01 crc kubenswrapper[4998]: I1203 16:25:01.098770 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:25:01 crc kubenswrapper[4998]: I1203 16:25:01.150473 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-57c877d99d-7qbcv" event={"ID":"050d52db-a22f-43de-bf07-ed0b6bed49db","Type":"ContainerStarted","Data":"691b1db2c406752340e3a55c14dc904ba161ca938c5b03710caf8aae59494ef6"} Dec 03 16:25:01 crc kubenswrapper[4998]: I1203 16:25:01.150538 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-57c877d99d-7qbcv" event={"ID":"050d52db-a22f-43de-bf07-ed0b6bed49db","Type":"ContainerStarted","Data":"2bd2dfa9718f409fa5de818747f3ac182a76cfcbd66cb8dd5f39434e965caf0f"} Dec 03 16:25:01 crc kubenswrapper[4998]: I1203 16:25:01.155206 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"a009aeee-c849-4c00-b41b-3cf5e36a75bb","Type":"ContainerStarted","Data":"4b34811f72882bb0f24e132377974281dacd94e6a42158df320e63964f1961b2"} Dec 03 16:25:01 crc kubenswrapper[4998]: I1203 16:25:01.190722 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=2.19070277 podStartE2EDuration="2.19070277s" podCreationTimestamp="2025-12-03 16:24:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:25:01.176210728 +0000 UTC m=+1279.787910951" watchObservedRunningTime="2025-12-03 16:25:01.19070277 +0000 UTC m=+1279.802402993" Dec 03 16:25:01 crc kubenswrapper[4998]: I1203 16:25:01.276041 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7c5f7f9446-f97x8" Dec 03 16:25:01 crc kubenswrapper[4998]: I1203 16:25:01.340502 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7565d78f8d-jm9qt"] Dec 03 16:25:01 crc kubenswrapper[4998]: I1203 16:25:01.340751 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7565d78f8d-jm9qt" podUID="d7600639-7175-4c0a-9524-40adb505d3ea" containerName="horizon-log" containerID="cri-o://75a61917c7bd2bcb544cb164de1c7e5b8e8e2cbe5262c7fa8075c0db387aa835" gracePeriod=30 Dec 03 16:25:01 crc kubenswrapper[4998]: I1203 16:25:01.340916 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7565d78f8d-jm9qt" podUID="d7600639-7175-4c0a-9524-40adb505d3ea" containerName="horizon" containerID="cri-o://549d68f35b182d6fa1f0327a26451914c3855dd97314b4621e055785c688e514" gracePeriod=30 Dec 03 16:25:02 crc kubenswrapper[4998]: E1203 16:25:02.463137 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2c4f81d00299bd0d05b7f51e1c0c848a779a1b0b6f199befff2e205150316838" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 16:25:02 crc kubenswrapper[4998]: E1203 16:25:02.464742 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2c4f81d00299bd0d05b7f51e1c0c848a779a1b0b6f199befff2e205150316838" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 16:25:02 crc kubenswrapper[4998]: E1203 16:25:02.466151 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2c4f81d00299bd0d05b7f51e1c0c848a779a1b0b6f199befff2e205150316838" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 16:25:02 crc kubenswrapper[4998]: E1203 16:25:02.466181 4998 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/watcher-applier-0" podUID="5b94f565-0948-4e4f-9636-2833d43bb829" containerName="watcher-applier" Dec 03 16:25:03 crc kubenswrapper[4998]: I1203 16:25:03.178783 4998 generic.go:334] "Generic (PLEG): container finished" podID="d7600639-7175-4c0a-9524-40adb505d3ea" containerID="549d68f35b182d6fa1f0327a26451914c3855dd97314b4621e055785c688e514" exitCode=0 Dec 03 16:25:03 crc kubenswrapper[4998]: I1203 16:25:03.178827 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7565d78f8d-jm9qt" event={"ID":"d7600639-7175-4c0a-9524-40adb505d3ea","Type":"ContainerDied","Data":"549d68f35b182d6fa1f0327a26451914c3855dd97314b4621e055785c688e514"} Dec 03 16:25:04 crc kubenswrapper[4998]: I1203 16:25:04.191372 4998 generic.go:334] "Generic (PLEG): container finished" podID="a009aeee-c849-4c00-b41b-3cf5e36a75bb" containerID="4b34811f72882bb0f24e132377974281dacd94e6a42158df320e63964f1961b2" exitCode=1 Dec 03 16:25:04 crc kubenswrapper[4998]: I1203 16:25:04.191413 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"a009aeee-c849-4c00-b41b-3cf5e36a75bb","Type":"ContainerDied","Data":"4b34811f72882bb0f24e132377974281dacd94e6a42158df320e63964f1961b2"} Dec 03 16:25:04 crc kubenswrapper[4998]: I1203 16:25:04.192179 4998 scope.go:117] "RemoveContainer" containerID="4b34811f72882bb0f24e132377974281dacd94e6a42158df320e63964f1961b2" Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.203560 4998 generic.go:334] "Generic (PLEG): container finished" podID="ef216407-b6ec-4691-8f49-cf8fce3f3544" containerID="d4b7cb504608696bdfe943b2476e64150c5012ce77fbcbf7167ba50a7e549607" exitCode=137 Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.203597 4998 generic.go:334] "Generic (PLEG): container finished" podID="ef216407-b6ec-4691-8f49-cf8fce3f3544" containerID="acfefbd1e9cf3e1a042744e581a4aa19f86d8779dde8da25311f8659c3f3bd73" exitCode=137 Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.203657 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67bd586fc-x2g2p" event={"ID":"ef216407-b6ec-4691-8f49-cf8fce3f3544","Type":"ContainerDied","Data":"d4b7cb504608696bdfe943b2476e64150c5012ce77fbcbf7167ba50a7e549607"} Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.203719 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67bd586fc-x2g2p" event={"ID":"ef216407-b6ec-4691-8f49-cf8fce3f3544","Type":"ContainerDied","Data":"acfefbd1e9cf3e1a042744e581a4aa19f86d8779dde8da25311f8659c3f3bd73"} Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.206060 4998 generic.go:334] "Generic (PLEG): container finished" podID="53f87c55-ecee-4d26-8244-6181b21dece0" containerID="72038f1b94d45b72f211d0025db98351b1b4c797dfa770ea034904cf1c1b58ff" exitCode=137 Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.206103 4998 generic.go:334] "Generic (PLEG): container finished" podID="53f87c55-ecee-4d26-8244-6181b21dece0" containerID="95037244abd2cb3d3c77b85237199f52984305069d5d070b3c10e0a4cdefeab5" exitCode=137 Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.206146 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-648fb89557-mvw59" event={"ID":"53f87c55-ecee-4d26-8244-6181b21dece0","Type":"ContainerDied","Data":"72038f1b94d45b72f211d0025db98351b1b4c797dfa770ea034904cf1c1b58ff"} Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.206237 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-648fb89557-mvw59" event={"ID":"53f87c55-ecee-4d26-8244-6181b21dece0","Type":"ContainerDied","Data":"95037244abd2cb3d3c77b85237199f52984305069d5d070b3c10e0a4cdefeab5"} Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.624612 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-648fb89557-mvw59" Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.676302 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67bd586fc-x2g2p" Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.716539 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2x5cz\" (UniqueName: \"kubernetes.io/projected/53f87c55-ecee-4d26-8244-6181b21dece0-kube-api-access-2x5cz\") pod \"53f87c55-ecee-4d26-8244-6181b21dece0\" (UID: \"53f87c55-ecee-4d26-8244-6181b21dece0\") " Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.716638 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53f87c55-ecee-4d26-8244-6181b21dece0-logs\") pod \"53f87c55-ecee-4d26-8244-6181b21dece0\" (UID: \"53f87c55-ecee-4d26-8244-6181b21dece0\") " Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.716709 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53f87c55-ecee-4d26-8244-6181b21dece0-config-data\") pod \"53f87c55-ecee-4d26-8244-6181b21dece0\" (UID: \"53f87c55-ecee-4d26-8244-6181b21dece0\") " Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.716772 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53f87c55-ecee-4d26-8244-6181b21dece0-scripts\") pod \"53f87c55-ecee-4d26-8244-6181b21dece0\" (UID: \"53f87c55-ecee-4d26-8244-6181b21dece0\") " Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.717170 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/53f87c55-ecee-4d26-8244-6181b21dece0-horizon-secret-key\") pod \"53f87c55-ecee-4d26-8244-6181b21dece0\" (UID: \"53f87c55-ecee-4d26-8244-6181b21dece0\") " Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.717208 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53f87c55-ecee-4d26-8244-6181b21dece0-logs" (OuterVolumeSpecName: "logs") pod "53f87c55-ecee-4d26-8244-6181b21dece0" (UID: "53f87c55-ecee-4d26-8244-6181b21dece0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.717840 4998 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53f87c55-ecee-4d26-8244-6181b21dece0-logs\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.721381 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53f87c55-ecee-4d26-8244-6181b21dece0-kube-api-access-2x5cz" (OuterVolumeSpecName: "kube-api-access-2x5cz") pod "53f87c55-ecee-4d26-8244-6181b21dece0" (UID: "53f87c55-ecee-4d26-8244-6181b21dece0"). InnerVolumeSpecName "kube-api-access-2x5cz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.721830 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53f87c55-ecee-4d26-8244-6181b21dece0-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "53f87c55-ecee-4d26-8244-6181b21dece0" (UID: "53f87c55-ecee-4d26-8244-6181b21dece0"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.739280 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53f87c55-ecee-4d26-8244-6181b21dece0-scripts" (OuterVolumeSpecName: "scripts") pod "53f87c55-ecee-4d26-8244-6181b21dece0" (UID: "53f87c55-ecee-4d26-8244-6181b21dece0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.745456 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53f87c55-ecee-4d26-8244-6181b21dece0-config-data" (OuterVolumeSpecName: "config-data") pod "53f87c55-ecee-4d26-8244-6181b21dece0" (UID: "53f87c55-ecee-4d26-8244-6181b21dece0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.818937 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ef216407-b6ec-4691-8f49-cf8fce3f3544-horizon-secret-key\") pod \"ef216407-b6ec-4691-8f49-cf8fce3f3544\" (UID: \"ef216407-b6ec-4691-8f49-cf8fce3f3544\") " Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.818991 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef216407-b6ec-4691-8f49-cf8fce3f3544-logs\") pod \"ef216407-b6ec-4691-8f49-cf8fce3f3544\" (UID: \"ef216407-b6ec-4691-8f49-cf8fce3f3544\") " Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.819058 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlq82\" (UniqueName: \"kubernetes.io/projected/ef216407-b6ec-4691-8f49-cf8fce3f3544-kube-api-access-rlq82\") pod \"ef216407-b6ec-4691-8f49-cf8fce3f3544\" (UID: \"ef216407-b6ec-4691-8f49-cf8fce3f3544\") " Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.819130 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef216407-b6ec-4691-8f49-cf8fce3f3544-config-data\") pod \"ef216407-b6ec-4691-8f49-cf8fce3f3544\" (UID: \"ef216407-b6ec-4691-8f49-cf8fce3f3544\") " Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.819268 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef216407-b6ec-4691-8f49-cf8fce3f3544-scripts\") pod \"ef216407-b6ec-4691-8f49-cf8fce3f3544\" (UID: \"ef216407-b6ec-4691-8f49-cf8fce3f3544\") " Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.819697 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2x5cz\" (UniqueName: \"kubernetes.io/projected/53f87c55-ecee-4d26-8244-6181b21dece0-kube-api-access-2x5cz\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.819715 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53f87c55-ecee-4d26-8244-6181b21dece0-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.819724 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53f87c55-ecee-4d26-8244-6181b21dece0-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.819732 4998 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/53f87c55-ecee-4d26-8244-6181b21dece0-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.820562 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef216407-b6ec-4691-8f49-cf8fce3f3544-logs" (OuterVolumeSpecName: "logs") pod "ef216407-b6ec-4691-8f49-cf8fce3f3544" (UID: "ef216407-b6ec-4691-8f49-cf8fce3f3544"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.822433 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef216407-b6ec-4691-8f49-cf8fce3f3544-kube-api-access-rlq82" (OuterVolumeSpecName: "kube-api-access-rlq82") pod "ef216407-b6ec-4691-8f49-cf8fce3f3544" (UID: "ef216407-b6ec-4691-8f49-cf8fce3f3544"). InnerVolumeSpecName "kube-api-access-rlq82". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.822747 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef216407-b6ec-4691-8f49-cf8fce3f3544-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "ef216407-b6ec-4691-8f49-cf8fce3f3544" (UID: "ef216407-b6ec-4691-8f49-cf8fce3f3544"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.840411 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef216407-b6ec-4691-8f49-cf8fce3f3544-scripts" (OuterVolumeSpecName: "scripts") pod "ef216407-b6ec-4691-8f49-cf8fce3f3544" (UID: "ef216407-b6ec-4691-8f49-cf8fce3f3544"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.859897 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef216407-b6ec-4691-8f49-cf8fce3f3544-config-data" (OuterVolumeSpecName: "config-data") pod "ef216407-b6ec-4691-8f49-cf8fce3f3544" (UID: "ef216407-b6ec-4691-8f49-cf8fce3f3544"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.921736 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef216407-b6ec-4691-8f49-cf8fce3f3544-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.921784 4998 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ef216407-b6ec-4691-8f49-cf8fce3f3544-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.921796 4998 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef216407-b6ec-4691-8f49-cf8fce3f3544-logs\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.921804 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlq82\" (UniqueName: \"kubernetes.io/projected/ef216407-b6ec-4691-8f49-cf8fce3f3544-kube-api-access-rlq82\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:05 crc kubenswrapper[4998]: I1203 16:25:05.921812 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef216407-b6ec-4691-8f49-cf8fce3f3544-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:06 crc kubenswrapper[4998]: I1203 16:25:06.225529 4998 generic.go:334] "Generic (PLEG): container finished" podID="5d2429d8-9baa-4532-9f8b-9ffb72a3a764" containerID="9c2779da866bb38d647d67f9337f9006e0141dc294f2b196b28cf1e700822fcf" exitCode=0 Dec 03 16:25:06 crc kubenswrapper[4998]: I1203 16:25:06.225661 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-w8k5f" event={"ID":"5d2429d8-9baa-4532-9f8b-9ffb72a3a764","Type":"ContainerDied","Data":"9c2779da866bb38d647d67f9337f9006e0141dc294f2b196b28cf1e700822fcf"} Dec 03 16:25:06 crc kubenswrapper[4998]: I1203 16:25:06.229048 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4faa24a-a43f-4d24-b20d-09e241d0cc96","Type":"ContainerStarted","Data":"f4e16b7cd25c4006759bea8a0ecae336123d5c61dcd9a285eafc5e9515aa15d4"} Dec 03 16:25:06 crc kubenswrapper[4998]: I1203 16:25:06.232171 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67bd586fc-x2g2p" event={"ID":"ef216407-b6ec-4691-8f49-cf8fce3f3544","Type":"ContainerDied","Data":"683c721d22b4678bf597f95fb591e6a027d68e3052836353969ee8d7e7a863ba"} Dec 03 16:25:06 crc kubenswrapper[4998]: I1203 16:25:06.232232 4998 scope.go:117] "RemoveContainer" containerID="d4b7cb504608696bdfe943b2476e64150c5012ce77fbcbf7167ba50a7e549607" Dec 03 16:25:06 crc kubenswrapper[4998]: I1203 16:25:06.232423 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67bd586fc-x2g2p" Dec 03 16:25:06 crc kubenswrapper[4998]: I1203 16:25:06.253219 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-648fb89557-mvw59" event={"ID":"53f87c55-ecee-4d26-8244-6181b21dece0","Type":"ContainerDied","Data":"07e18fb60fe608ea212b8c98f530c643287fbdff81593c6a34fefa9c5b99e086"} Dec 03 16:25:06 crc kubenswrapper[4998]: I1203 16:25:06.253385 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-648fb89557-mvw59" Dec 03 16:25:06 crc kubenswrapper[4998]: I1203 16:25:06.286950 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-67bd586fc-x2g2p"] Dec 03 16:25:06 crc kubenswrapper[4998]: I1203 16:25:06.294858 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-67bd586fc-x2g2p"] Dec 03 16:25:06 crc kubenswrapper[4998]: I1203 16:25:06.310006 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-648fb89557-mvw59"] Dec 03 16:25:06 crc kubenswrapper[4998]: I1203 16:25:06.316732 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-648fb89557-mvw59"] Dec 03 16:25:06 crc kubenswrapper[4998]: I1203 16:25:06.439417 4998 scope.go:117] "RemoveContainer" containerID="acfefbd1e9cf3e1a042744e581a4aa19f86d8779dde8da25311f8659c3f3bd73" Dec 03 16:25:06 crc kubenswrapper[4998]: I1203 16:25:06.464585 4998 scope.go:117] "RemoveContainer" containerID="72038f1b94d45b72f211d0025db98351b1b4c797dfa770ea034904cf1c1b58ff" Dec 03 16:25:06 crc kubenswrapper[4998]: I1203 16:25:06.598614 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7565d78f8d-jm9qt" podUID="d7600639-7175-4c0a-9524-40adb505d3ea" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.159:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.159:8443: connect: connection refused" Dec 03 16:25:06 crc kubenswrapper[4998]: I1203 16:25:06.631869 4998 scope.go:117] "RemoveContainer" containerID="95037244abd2cb3d3c77b85237199f52984305069d5d070b3c10e0a4cdefeab5" Dec 03 16:25:07 crc kubenswrapper[4998]: E1203 16:25:07.465919 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2c4f81d00299bd0d05b7f51e1c0c848a779a1b0b6f199befff2e205150316838" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 16:25:07 crc kubenswrapper[4998]: E1203 16:25:07.468098 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2c4f81d00299bd0d05b7f51e1c0c848a779a1b0b6f199befff2e205150316838" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 16:25:07 crc kubenswrapper[4998]: E1203 16:25:07.469547 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2c4f81d00299bd0d05b7f51e1c0c848a779a1b0b6f199befff2e205150316838" cmd=["/usr/bin/pgrep","-r","DRST","watcher-applier"] Dec 03 16:25:07 crc kubenswrapper[4998]: E1203 16:25:07.469579 4998 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/watcher-applier-0" podUID="5b94f565-0948-4e4f-9636-2833d43bb829" containerName="watcher-applier" Dec 03 16:25:07 crc kubenswrapper[4998]: I1203 16:25:07.653547 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-w8k5f" Dec 03 16:25:07 crc kubenswrapper[4998]: I1203 16:25:07.690870 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53f87c55-ecee-4d26-8244-6181b21dece0" path="/var/lib/kubelet/pods/53f87c55-ecee-4d26-8244-6181b21dece0/volumes" Dec 03 16:25:07 crc kubenswrapper[4998]: I1203 16:25:07.691776 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef216407-b6ec-4691-8f49-cf8fce3f3544" path="/var/lib/kubelet/pods/ef216407-b6ec-4691-8f49-cf8fce3f3544/volumes" Dec 03 16:25:07 crc kubenswrapper[4998]: I1203 16:25:07.770666 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d2429d8-9baa-4532-9f8b-9ffb72a3a764-combined-ca-bundle\") pod \"5d2429d8-9baa-4532-9f8b-9ffb72a3a764\" (UID: \"5d2429d8-9baa-4532-9f8b-9ffb72a3a764\") " Dec 03 16:25:07 crc kubenswrapper[4998]: I1203 16:25:07.770748 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfq6t\" (UniqueName: \"kubernetes.io/projected/5d2429d8-9baa-4532-9f8b-9ffb72a3a764-kube-api-access-pfq6t\") pod \"5d2429d8-9baa-4532-9f8b-9ffb72a3a764\" (UID: \"5d2429d8-9baa-4532-9f8b-9ffb72a3a764\") " Dec 03 16:25:07 crc kubenswrapper[4998]: I1203 16:25:07.770823 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5d2429d8-9baa-4532-9f8b-9ffb72a3a764-db-sync-config-data\") pod \"5d2429d8-9baa-4532-9f8b-9ffb72a3a764\" (UID: \"5d2429d8-9baa-4532-9f8b-9ffb72a3a764\") " Dec 03 16:25:07 crc kubenswrapper[4998]: I1203 16:25:07.777673 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d2429d8-9baa-4532-9f8b-9ffb72a3a764-kube-api-access-pfq6t" (OuterVolumeSpecName: "kube-api-access-pfq6t") pod "5d2429d8-9baa-4532-9f8b-9ffb72a3a764" (UID: "5d2429d8-9baa-4532-9f8b-9ffb72a3a764"). InnerVolumeSpecName "kube-api-access-pfq6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:25:07 crc kubenswrapper[4998]: I1203 16:25:07.778020 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d2429d8-9baa-4532-9f8b-9ffb72a3a764-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "5d2429d8-9baa-4532-9f8b-9ffb72a3a764" (UID: "5d2429d8-9baa-4532-9f8b-9ffb72a3a764"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:07 crc kubenswrapper[4998]: I1203 16:25:07.807563 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d2429d8-9baa-4532-9f8b-9ffb72a3a764-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d2429d8-9baa-4532-9f8b-9ffb72a3a764" (UID: "5d2429d8-9baa-4532-9f8b-9ffb72a3a764"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:07 crc kubenswrapper[4998]: I1203 16:25:07.875053 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d2429d8-9baa-4532-9f8b-9ffb72a3a764-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:07 crc kubenswrapper[4998]: I1203 16:25:07.875114 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfq6t\" (UniqueName: \"kubernetes.io/projected/5d2429d8-9baa-4532-9f8b-9ffb72a3a764-kube-api-access-pfq6t\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:07 crc kubenswrapper[4998]: I1203 16:25:07.875138 4998 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5d2429d8-9baa-4532-9f8b-9ffb72a3a764-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.282014 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-57c877d99d-7qbcv" event={"ID":"050d52db-a22f-43de-bf07-ed0b6bed49db","Type":"ContainerStarted","Data":"748d95f94b0930fa869c3e01bacdc3c5cf1336444e4cf7cf06238e40f33954f1"} Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.283244 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.283818 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.290639 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"a009aeee-c849-4c00-b41b-3cf5e36a75bb","Type":"ContainerStarted","Data":"616cb19e04d79255ec317f93c152a8d4bf88d542e297bf72a69bea31d5e86255"} Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.294697 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e4faa24a-a43f-4d24-b20d-09e241d0cc96" containerName="ceilometer-central-agent" containerID="cri-o://5ae86f8014172d9b95a4494ff98b6cbc3dfc4a7399092fa9f3e71c9260a39e54" gracePeriod=30 Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.294941 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-w8k5f" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.294998 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e4faa24a-a43f-4d24-b20d-09e241d0cc96" containerName="proxy-httpd" containerID="cri-o://f4e16b7cd25c4006759bea8a0ecae336123d5c61dcd9a285eafc5e9515aa15d4" gracePeriod=30 Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.295063 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e4faa24a-a43f-4d24-b20d-09e241d0cc96" containerName="sg-core" containerID="cri-o://6f0d6533d024330f91e4fcc8a4ad9f21d7ed7fb31327d7ca16b483a8f88d0308" gracePeriod=30 Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.295105 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e4faa24a-a43f-4d24-b20d-09e241d0cc96" containerName="ceilometer-notification-agent" containerID="cri-o://a9100307a749acaf5cb3ca1c61dbd1c3014727d4fc73d649fe5f2633f22ad98e" gracePeriod=30 Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.295165 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-w8k5f" event={"ID":"5d2429d8-9baa-4532-9f8b-9ffb72a3a764","Type":"ContainerDied","Data":"f7b8c0dcf9cc5974b58aa07b9c714976da85b6a9e3adee1828e512e498b2c95f"} Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.295182 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7b8c0dcf9cc5974b58aa07b9c714976da85b6a9e3adee1828e512e498b2c95f" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.295196 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.329401 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-57c877d99d-7qbcv" podStartSLOduration=8.329384262 podStartE2EDuration="8.329384262s" podCreationTimestamp="2025-12-03 16:25:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:25:08.312707036 +0000 UTC m=+1286.924407259" watchObservedRunningTime="2025-12-03 16:25:08.329384262 +0000 UTC m=+1286.941084485" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.351888 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=5.178883393 podStartE2EDuration="1m11.351867879s" podCreationTimestamp="2025-12-03 16:23:57 +0000 UTC" firstStartedPulling="2025-12-03 16:23:59.283274166 +0000 UTC m=+1217.894974389" lastFinishedPulling="2025-12-03 16:25:05.456258662 +0000 UTC m=+1284.067958875" observedRunningTime="2025-12-03 16:25:08.335939821 +0000 UTC m=+1286.947640054" watchObservedRunningTime="2025-12-03 16:25:08.351867879 +0000 UTC m=+1286.963568102" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.524776 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7d588c5fc8-x85fx"] Dec 03 16:25:08 crc kubenswrapper[4998]: E1203 16:25:08.525276 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef216407-b6ec-4691-8f49-cf8fce3f3544" containerName="horizon-log" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.525294 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef216407-b6ec-4691-8f49-cf8fce3f3544" containerName="horizon-log" Dec 03 16:25:08 crc kubenswrapper[4998]: E1203 16:25:08.525307 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d2429d8-9baa-4532-9f8b-9ffb72a3a764" containerName="barbican-db-sync" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.525315 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d2429d8-9baa-4532-9f8b-9ffb72a3a764" containerName="barbican-db-sync" Dec 03 16:25:08 crc kubenswrapper[4998]: E1203 16:25:08.525340 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef216407-b6ec-4691-8f49-cf8fce3f3544" containerName="horizon" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.525351 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef216407-b6ec-4691-8f49-cf8fce3f3544" containerName="horizon" Dec 03 16:25:08 crc kubenswrapper[4998]: E1203 16:25:08.525381 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53f87c55-ecee-4d26-8244-6181b21dece0" containerName="horizon-log" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.525389 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="53f87c55-ecee-4d26-8244-6181b21dece0" containerName="horizon-log" Dec 03 16:25:08 crc kubenswrapper[4998]: E1203 16:25:08.525408 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53f87c55-ecee-4d26-8244-6181b21dece0" containerName="horizon" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.525416 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="53f87c55-ecee-4d26-8244-6181b21dece0" containerName="horizon" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.525612 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d2429d8-9baa-4532-9f8b-9ffb72a3a764" containerName="barbican-db-sync" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.525634 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="53f87c55-ecee-4d26-8244-6181b21dece0" containerName="horizon-log" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.525649 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="53f87c55-ecee-4d26-8244-6181b21dece0" containerName="horizon" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.525665 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef216407-b6ec-4691-8f49-cf8fce3f3544" containerName="horizon" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.525691 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef216407-b6ec-4691-8f49-cf8fce3f3544" containerName="horizon-log" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.526912 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7d588c5fc8-x85fx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.532200 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-clw2v" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.532309 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.532426 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.532854 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7d6d5bf649-h8gnx"] Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.535122 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7d6d5bf649-h8gnx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.548250 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.563844 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7d6d5bf649-h8gnx"] Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.574102 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7d588c5fc8-x85fx"] Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.585737 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57638b87-e1dd-4583-acc1-d57eb123de4d-combined-ca-bundle\") pod \"barbican-keystone-listener-7d588c5fc8-x85fx\" (UID: \"57638b87-e1dd-4583-acc1-d57eb123de4d\") " pod="openstack/barbican-keystone-listener-7d588c5fc8-x85fx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.585866 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57638b87-e1dd-4583-acc1-d57eb123de4d-logs\") pod \"barbican-keystone-listener-7d588c5fc8-x85fx\" (UID: \"57638b87-e1dd-4583-acc1-d57eb123de4d\") " pod="openstack/barbican-keystone-listener-7d588c5fc8-x85fx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.586002 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/57638b87-e1dd-4583-acc1-d57eb123de4d-config-data-custom\") pod \"barbican-keystone-listener-7d588c5fc8-x85fx\" (UID: \"57638b87-e1dd-4583-acc1-d57eb123de4d\") " pod="openstack/barbican-keystone-listener-7d588c5fc8-x85fx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.586088 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57638b87-e1dd-4583-acc1-d57eb123de4d-config-data\") pod \"barbican-keystone-listener-7d588c5fc8-x85fx\" (UID: \"57638b87-e1dd-4583-acc1-d57eb123de4d\") " pod="openstack/barbican-keystone-listener-7d588c5fc8-x85fx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.586161 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dt4l\" (UniqueName: \"kubernetes.io/projected/57638b87-e1dd-4583-acc1-d57eb123de4d-kube-api-access-2dt4l\") pod \"barbican-keystone-listener-7d588c5fc8-x85fx\" (UID: \"57638b87-e1dd-4583-acc1-d57eb123de4d\") " pod="openstack/barbican-keystone-listener-7d588c5fc8-x85fx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.687976 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-fcb95679f-wbxs5"] Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.689626 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.691528 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d33fe6f6-a02c-4841-baf4-1c956cd2e9f2-combined-ca-bundle\") pod \"barbican-worker-7d6d5bf649-h8gnx\" (UID: \"d33fe6f6-a02c-4841-baf4-1c956cd2e9f2\") " pod="openstack/barbican-worker-7d6d5bf649-h8gnx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.691605 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/57638b87-e1dd-4583-acc1-d57eb123de4d-config-data-custom\") pod \"barbican-keystone-listener-7d588c5fc8-x85fx\" (UID: \"57638b87-e1dd-4583-acc1-d57eb123de4d\") " pod="openstack/barbican-keystone-listener-7d588c5fc8-x85fx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.691629 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57638b87-e1dd-4583-acc1-d57eb123de4d-config-data\") pod \"barbican-keystone-listener-7d588c5fc8-x85fx\" (UID: \"57638b87-e1dd-4583-acc1-d57eb123de4d\") " pod="openstack/barbican-keystone-listener-7d588c5fc8-x85fx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.691656 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dt4l\" (UniqueName: \"kubernetes.io/projected/57638b87-e1dd-4583-acc1-d57eb123de4d-kube-api-access-2dt4l\") pod \"barbican-keystone-listener-7d588c5fc8-x85fx\" (UID: \"57638b87-e1dd-4583-acc1-d57eb123de4d\") " pod="openstack/barbican-keystone-listener-7d588c5fc8-x85fx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.691679 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d33fe6f6-a02c-4841-baf4-1c956cd2e9f2-config-data-custom\") pod \"barbican-worker-7d6d5bf649-h8gnx\" (UID: \"d33fe6f6-a02c-4841-baf4-1c956cd2e9f2\") " pod="openstack/barbican-worker-7d6d5bf649-h8gnx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.691700 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d33fe6f6-a02c-4841-baf4-1c956cd2e9f2-logs\") pod \"barbican-worker-7d6d5bf649-h8gnx\" (UID: \"d33fe6f6-a02c-4841-baf4-1c956cd2e9f2\") " pod="openstack/barbican-worker-7d6d5bf649-h8gnx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.691720 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2krf\" (UniqueName: \"kubernetes.io/projected/d33fe6f6-a02c-4841-baf4-1c956cd2e9f2-kube-api-access-n2krf\") pod \"barbican-worker-7d6d5bf649-h8gnx\" (UID: \"d33fe6f6-a02c-4841-baf4-1c956cd2e9f2\") " pod="openstack/barbican-worker-7d6d5bf649-h8gnx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.691813 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57638b87-e1dd-4583-acc1-d57eb123de4d-combined-ca-bundle\") pod \"barbican-keystone-listener-7d588c5fc8-x85fx\" (UID: \"57638b87-e1dd-4583-acc1-d57eb123de4d\") " pod="openstack/barbican-keystone-listener-7d588c5fc8-x85fx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.691860 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57638b87-e1dd-4583-acc1-d57eb123de4d-logs\") pod \"barbican-keystone-listener-7d588c5fc8-x85fx\" (UID: \"57638b87-e1dd-4583-acc1-d57eb123de4d\") " pod="openstack/barbican-keystone-listener-7d588c5fc8-x85fx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.691887 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d33fe6f6-a02c-4841-baf4-1c956cd2e9f2-config-data\") pod \"barbican-worker-7d6d5bf649-h8gnx\" (UID: \"d33fe6f6-a02c-4841-baf4-1c956cd2e9f2\") " pod="openstack/barbican-worker-7d6d5bf649-h8gnx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.693775 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/57638b87-e1dd-4583-acc1-d57eb123de4d-logs\") pod \"barbican-keystone-listener-7d588c5fc8-x85fx\" (UID: \"57638b87-e1dd-4583-acc1-d57eb123de4d\") " pod="openstack/barbican-keystone-listener-7d588c5fc8-x85fx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.704897 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57638b87-e1dd-4583-acc1-d57eb123de4d-combined-ca-bundle\") pod \"barbican-keystone-listener-7d588c5fc8-x85fx\" (UID: \"57638b87-e1dd-4583-acc1-d57eb123de4d\") " pod="openstack/barbican-keystone-listener-7d588c5fc8-x85fx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.706026 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/57638b87-e1dd-4583-acc1-d57eb123de4d-config-data-custom\") pod \"barbican-keystone-listener-7d588c5fc8-x85fx\" (UID: \"57638b87-e1dd-4583-acc1-d57eb123de4d\") " pod="openstack/barbican-keystone-listener-7d588c5fc8-x85fx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.732789 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57638b87-e1dd-4583-acc1-d57eb123de4d-config-data\") pod \"barbican-keystone-listener-7d588c5fc8-x85fx\" (UID: \"57638b87-e1dd-4583-acc1-d57eb123de4d\") " pod="openstack/barbican-keystone-listener-7d588c5fc8-x85fx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.743344 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dt4l\" (UniqueName: \"kubernetes.io/projected/57638b87-e1dd-4583-acc1-d57eb123de4d-kube-api-access-2dt4l\") pod \"barbican-keystone-listener-7d588c5fc8-x85fx\" (UID: \"57638b87-e1dd-4583-acc1-d57eb123de4d\") " pod="openstack/barbican-keystone-listener-7d588c5fc8-x85fx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.775824 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fcb95679f-wbxs5"] Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.797644 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-ovsdbserver-nb\") pod \"dnsmasq-dns-fcb95679f-wbxs5\" (UID: \"c61c77db-8343-4090-ad97-06d7814f475c\") " pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.797717 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d33fe6f6-a02c-4841-baf4-1c956cd2e9f2-config-data-custom\") pod \"barbican-worker-7d6d5bf649-h8gnx\" (UID: \"d33fe6f6-a02c-4841-baf4-1c956cd2e9f2\") " pod="openstack/barbican-worker-7d6d5bf649-h8gnx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.797739 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d33fe6f6-a02c-4841-baf4-1c956cd2e9f2-logs\") pod \"barbican-worker-7d6d5bf649-h8gnx\" (UID: \"d33fe6f6-a02c-4841-baf4-1c956cd2e9f2\") " pod="openstack/barbican-worker-7d6d5bf649-h8gnx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.797770 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2krf\" (UniqueName: \"kubernetes.io/projected/d33fe6f6-a02c-4841-baf4-1c956cd2e9f2-kube-api-access-n2krf\") pod \"barbican-worker-7d6d5bf649-h8gnx\" (UID: \"d33fe6f6-a02c-4841-baf4-1c956cd2e9f2\") " pod="openstack/barbican-worker-7d6d5bf649-h8gnx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.797790 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c7t6\" (UniqueName: \"kubernetes.io/projected/c61c77db-8343-4090-ad97-06d7814f475c-kube-api-access-9c7t6\") pod \"dnsmasq-dns-fcb95679f-wbxs5\" (UID: \"c61c77db-8343-4090-ad97-06d7814f475c\") " pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.797839 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-ovsdbserver-sb\") pod \"dnsmasq-dns-fcb95679f-wbxs5\" (UID: \"c61c77db-8343-4090-ad97-06d7814f475c\") " pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.797858 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-config\") pod \"dnsmasq-dns-fcb95679f-wbxs5\" (UID: \"c61c77db-8343-4090-ad97-06d7814f475c\") " pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.797896 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-dns-svc\") pod \"dnsmasq-dns-fcb95679f-wbxs5\" (UID: \"c61c77db-8343-4090-ad97-06d7814f475c\") " pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.797916 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d33fe6f6-a02c-4841-baf4-1c956cd2e9f2-config-data\") pod \"barbican-worker-7d6d5bf649-h8gnx\" (UID: \"d33fe6f6-a02c-4841-baf4-1c956cd2e9f2\") " pod="openstack/barbican-worker-7d6d5bf649-h8gnx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.797948 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d33fe6f6-a02c-4841-baf4-1c956cd2e9f2-combined-ca-bundle\") pod \"barbican-worker-7d6d5bf649-h8gnx\" (UID: \"d33fe6f6-a02c-4841-baf4-1c956cd2e9f2\") " pod="openstack/barbican-worker-7d6d5bf649-h8gnx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.797971 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-dns-swift-storage-0\") pod \"dnsmasq-dns-fcb95679f-wbxs5\" (UID: \"c61c77db-8343-4090-ad97-06d7814f475c\") " pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.800112 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d33fe6f6-a02c-4841-baf4-1c956cd2e9f2-logs\") pod \"barbican-worker-7d6d5bf649-h8gnx\" (UID: \"d33fe6f6-a02c-4841-baf4-1c956cd2e9f2\") " pod="openstack/barbican-worker-7d6d5bf649-h8gnx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.803278 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d33fe6f6-a02c-4841-baf4-1c956cd2e9f2-config-data-custom\") pod \"barbican-worker-7d6d5bf649-h8gnx\" (UID: \"d33fe6f6-a02c-4841-baf4-1c956cd2e9f2\") " pod="openstack/barbican-worker-7d6d5bf649-h8gnx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.804086 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d33fe6f6-a02c-4841-baf4-1c956cd2e9f2-config-data\") pod \"barbican-worker-7d6d5bf649-h8gnx\" (UID: \"d33fe6f6-a02c-4841-baf4-1c956cd2e9f2\") " pod="openstack/barbican-worker-7d6d5bf649-h8gnx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.811473 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d33fe6f6-a02c-4841-baf4-1c956cd2e9f2-combined-ca-bundle\") pod \"barbican-worker-7d6d5bf649-h8gnx\" (UID: \"d33fe6f6-a02c-4841-baf4-1c956cd2e9f2\") " pod="openstack/barbican-worker-7d6d5bf649-h8gnx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.855313 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2krf\" (UniqueName: \"kubernetes.io/projected/d33fe6f6-a02c-4841-baf4-1c956cd2e9f2-kube-api-access-n2krf\") pod \"barbican-worker-7d6d5bf649-h8gnx\" (UID: \"d33fe6f6-a02c-4841-baf4-1c956cd2e9f2\") " pod="openstack/barbican-worker-7d6d5bf649-h8gnx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.899290 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-dns-swift-storage-0\") pod \"dnsmasq-dns-fcb95679f-wbxs5\" (UID: \"c61c77db-8343-4090-ad97-06d7814f475c\") " pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.899356 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-ovsdbserver-nb\") pod \"dnsmasq-dns-fcb95679f-wbxs5\" (UID: \"c61c77db-8343-4090-ad97-06d7814f475c\") " pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.899436 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c7t6\" (UniqueName: \"kubernetes.io/projected/c61c77db-8343-4090-ad97-06d7814f475c-kube-api-access-9c7t6\") pod \"dnsmasq-dns-fcb95679f-wbxs5\" (UID: \"c61c77db-8343-4090-ad97-06d7814f475c\") " pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.899505 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-ovsdbserver-sb\") pod \"dnsmasq-dns-fcb95679f-wbxs5\" (UID: \"c61c77db-8343-4090-ad97-06d7814f475c\") " pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.899540 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-config\") pod \"dnsmasq-dns-fcb95679f-wbxs5\" (UID: \"c61c77db-8343-4090-ad97-06d7814f475c\") " pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.899610 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-dns-svc\") pod \"dnsmasq-dns-fcb95679f-wbxs5\" (UID: \"c61c77db-8343-4090-ad97-06d7814f475c\") " pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.900800 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-dns-svc\") pod \"dnsmasq-dns-fcb95679f-wbxs5\" (UID: \"c61c77db-8343-4090-ad97-06d7814f475c\") " pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.903818 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-dns-swift-storage-0\") pod \"dnsmasq-dns-fcb95679f-wbxs5\" (UID: \"c61c77db-8343-4090-ad97-06d7814f475c\") " pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.906349 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-ovsdbserver-nb\") pod \"dnsmasq-dns-fcb95679f-wbxs5\" (UID: \"c61c77db-8343-4090-ad97-06d7814f475c\") " pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.909369 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6768bc7bbd-hsnsm"] Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.910867 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-ovsdbserver-sb\") pod \"dnsmasq-dns-fcb95679f-wbxs5\" (UID: \"c61c77db-8343-4090-ad97-06d7814f475c\") " pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.911147 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6768bc7bbd-hsnsm" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.911927 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7d588c5fc8-x85fx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.912477 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-config\") pod \"dnsmasq-dns-fcb95679f-wbxs5\" (UID: \"c61c77db-8343-4090-ad97-06d7814f475c\") " pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.913480 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.920141 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7d6d5bf649-h8gnx" Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.946473 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6768bc7bbd-hsnsm"] Dec 03 16:25:08 crc kubenswrapper[4998]: I1203 16:25:08.947729 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c7t6\" (UniqueName: \"kubernetes.io/projected/c61c77db-8343-4090-ad97-06d7814f475c-kube-api-access-9c7t6\") pod \"dnsmasq-dns-fcb95679f-wbxs5\" (UID: \"c61c77db-8343-4090-ad97-06d7814f475c\") " pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.003833 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q69g9\" (UniqueName: \"kubernetes.io/projected/fda7bb47-a754-4123-b031-14030f822167-kube-api-access-q69g9\") pod \"barbican-api-6768bc7bbd-hsnsm\" (UID: \"fda7bb47-a754-4123-b031-14030f822167\") " pod="openstack/barbican-api-6768bc7bbd-hsnsm" Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.003911 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fda7bb47-a754-4123-b031-14030f822167-logs\") pod \"barbican-api-6768bc7bbd-hsnsm\" (UID: \"fda7bb47-a754-4123-b031-14030f822167\") " pod="openstack/barbican-api-6768bc7bbd-hsnsm" Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.004001 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fda7bb47-a754-4123-b031-14030f822167-config-data-custom\") pod \"barbican-api-6768bc7bbd-hsnsm\" (UID: \"fda7bb47-a754-4123-b031-14030f822167\") " pod="openstack/barbican-api-6768bc7bbd-hsnsm" Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.004052 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fda7bb47-a754-4123-b031-14030f822167-config-data\") pod \"barbican-api-6768bc7bbd-hsnsm\" (UID: \"fda7bb47-a754-4123-b031-14030f822167\") " pod="openstack/barbican-api-6768bc7bbd-hsnsm" Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.004071 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fda7bb47-a754-4123-b031-14030f822167-combined-ca-bundle\") pod \"barbican-api-6768bc7bbd-hsnsm\" (UID: \"fda7bb47-a754-4123-b031-14030f822167\") " pod="openstack/barbican-api-6768bc7bbd-hsnsm" Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.108561 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fda7bb47-a754-4123-b031-14030f822167-config-data-custom\") pod \"barbican-api-6768bc7bbd-hsnsm\" (UID: \"fda7bb47-a754-4123-b031-14030f822167\") " pod="openstack/barbican-api-6768bc7bbd-hsnsm" Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.108941 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fda7bb47-a754-4123-b031-14030f822167-config-data\") pod \"barbican-api-6768bc7bbd-hsnsm\" (UID: \"fda7bb47-a754-4123-b031-14030f822167\") " pod="openstack/barbican-api-6768bc7bbd-hsnsm" Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.108959 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fda7bb47-a754-4123-b031-14030f822167-combined-ca-bundle\") pod \"barbican-api-6768bc7bbd-hsnsm\" (UID: \"fda7bb47-a754-4123-b031-14030f822167\") " pod="openstack/barbican-api-6768bc7bbd-hsnsm" Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.109035 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q69g9\" (UniqueName: \"kubernetes.io/projected/fda7bb47-a754-4123-b031-14030f822167-kube-api-access-q69g9\") pod \"barbican-api-6768bc7bbd-hsnsm\" (UID: \"fda7bb47-a754-4123-b031-14030f822167\") " pod="openstack/barbican-api-6768bc7bbd-hsnsm" Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.109068 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fda7bb47-a754-4123-b031-14030f822167-logs\") pod \"barbican-api-6768bc7bbd-hsnsm\" (UID: \"fda7bb47-a754-4123-b031-14030f822167\") " pod="openstack/barbican-api-6768bc7bbd-hsnsm" Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.109646 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fda7bb47-a754-4123-b031-14030f822167-logs\") pod \"barbican-api-6768bc7bbd-hsnsm\" (UID: \"fda7bb47-a754-4123-b031-14030f822167\") " pod="openstack/barbican-api-6768bc7bbd-hsnsm" Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.114016 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fda7bb47-a754-4123-b031-14030f822167-config-data-custom\") pod \"barbican-api-6768bc7bbd-hsnsm\" (UID: \"fda7bb47-a754-4123-b031-14030f822167\") " pod="openstack/barbican-api-6768bc7bbd-hsnsm" Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.115613 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fda7bb47-a754-4123-b031-14030f822167-config-data\") pod \"barbican-api-6768bc7bbd-hsnsm\" (UID: \"fda7bb47-a754-4123-b031-14030f822167\") " pod="openstack/barbican-api-6768bc7bbd-hsnsm" Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.118420 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fda7bb47-a754-4123-b031-14030f822167-combined-ca-bundle\") pod \"barbican-api-6768bc7bbd-hsnsm\" (UID: \"fda7bb47-a754-4123-b031-14030f822167\") " pod="openstack/barbican-api-6768bc7bbd-hsnsm" Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.132441 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q69g9\" (UniqueName: \"kubernetes.io/projected/fda7bb47-a754-4123-b031-14030f822167-kube-api-access-q69g9\") pod \"barbican-api-6768bc7bbd-hsnsm\" (UID: \"fda7bb47-a754-4123-b031-14030f822167\") " pod="openstack/barbican-api-6768bc7bbd-hsnsm" Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.239889 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.259230 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6768bc7bbd-hsnsm" Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.345876 4998 generic.go:334] "Generic (PLEG): container finished" podID="e4faa24a-a43f-4d24-b20d-09e241d0cc96" containerID="f4e16b7cd25c4006759bea8a0ecae336123d5c61dcd9a285eafc5e9515aa15d4" exitCode=0 Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.345904 4998 generic.go:334] "Generic (PLEG): container finished" podID="e4faa24a-a43f-4d24-b20d-09e241d0cc96" containerID="6f0d6533d024330f91e4fcc8a4ad9f21d7ed7fb31327d7ca16b483a8f88d0308" exitCode=2 Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.345911 4998 generic.go:334] "Generic (PLEG): container finished" podID="e4faa24a-a43f-4d24-b20d-09e241d0cc96" containerID="5ae86f8014172d9b95a4494ff98b6cbc3dfc4a7399092fa9f3e71c9260a39e54" exitCode=0 Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.346797 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4faa24a-a43f-4d24-b20d-09e241d0cc96","Type":"ContainerDied","Data":"f4e16b7cd25c4006759bea8a0ecae336123d5c61dcd9a285eafc5e9515aa15d4"} Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.346820 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4faa24a-a43f-4d24-b20d-09e241d0cc96","Type":"ContainerDied","Data":"6f0d6533d024330f91e4fcc8a4ad9f21d7ed7fb31327d7ca16b483a8f88d0308"} Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.346833 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4faa24a-a43f-4d24-b20d-09e241d0cc96","Type":"ContainerDied","Data":"5ae86f8014172d9b95a4494ff98b6cbc3dfc4a7399092fa9f3e71c9260a39e54"} Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.514774 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.548437 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7d6d5bf649-h8gnx"] Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.553940 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7d588c5fc8-x85fx"] Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.555428 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.593490 4998 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.771393 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fcb95679f-wbxs5"] Dec 03 16:25:09 crc kubenswrapper[4998]: W1203 16:25:09.775441 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc61c77db_8343_4090_ad97_06d7814f475c.slice/crio-50d1c77b3705953dd177fbf55913122d319a25fa662e9b970f3c4201d941e15e WatchSource:0}: Error finding container 50d1c77b3705953dd177fbf55913122d319a25fa662e9b970f3c4201d941e15e: Status 404 returned error can't find the container with id 50d1c77b3705953dd177fbf55913122d319a25fa662e9b970f3c4201d941e15e Dec 03 16:25:09 crc kubenswrapper[4998]: I1203 16:25:09.869091 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6768bc7bbd-hsnsm"] Dec 03 16:25:09 crc kubenswrapper[4998]: W1203 16:25:09.874955 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfda7bb47_a754_4123_b031_14030f822167.slice/crio-adcef4eee1ecba4ee0f6fee20c3ab70f63aa6a088c870b071fe2569666a4e7a1 WatchSource:0}: Error finding container adcef4eee1ecba4ee0f6fee20c3ab70f63aa6a088c870b071fe2569666a4e7a1: Status 404 returned error can't find the container with id adcef4eee1ecba4ee0f6fee20c3ab70f63aa6a088c870b071fe2569666a4e7a1 Dec 03 16:25:10 crc kubenswrapper[4998]: I1203 16:25:10.362098 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:10 crc kubenswrapper[4998]: I1203 16:25:10.362463 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7d588c5fc8-x85fx" event={"ID":"57638b87-e1dd-4583-acc1-d57eb123de4d","Type":"ContainerStarted","Data":"3d615091304128b56515ea9daabd00bc1ed83068f5a3ecca8788f168a4582b40"} Dec 03 16:25:10 crc kubenswrapper[4998]: I1203 16:25:10.369075 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6768bc7bbd-hsnsm" event={"ID":"fda7bb47-a754-4123-b031-14030f822167","Type":"ContainerStarted","Data":"9d3f8d796e12715831aedb33ba672326901694d1c6e8c35f8ebcf992ac8abcdd"} Dec 03 16:25:10 crc kubenswrapper[4998]: I1203 16:25:10.369120 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6768bc7bbd-hsnsm" event={"ID":"fda7bb47-a754-4123-b031-14030f822167","Type":"ContainerStarted","Data":"adcef4eee1ecba4ee0f6fee20c3ab70f63aa6a088c870b071fe2569666a4e7a1"} Dec 03 16:25:10 crc kubenswrapper[4998]: I1203 16:25:10.371877 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7d6d5bf649-h8gnx" event={"ID":"d33fe6f6-a02c-4841-baf4-1c956cd2e9f2","Type":"ContainerStarted","Data":"3e66bbf6b1621173e1adbb39c73a3a408b0e59c1eb7e2a90c86b8970b6936ce7"} Dec 03 16:25:10 crc kubenswrapper[4998]: I1203 16:25:10.376552 4998 generic.go:334] "Generic (PLEG): container finished" podID="5b94f565-0948-4e4f-9636-2833d43bb829" containerID="2c4f81d00299bd0d05b7f51e1c0c848a779a1b0b6f199befff2e205150316838" exitCode=137 Dec 03 16:25:10 crc kubenswrapper[4998]: I1203 16:25:10.376609 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"5b94f565-0948-4e4f-9636-2833d43bb829","Type":"ContainerDied","Data":"2c4f81d00299bd0d05b7f51e1c0c848a779a1b0b6f199befff2e205150316838"} Dec 03 16:25:10 crc kubenswrapper[4998]: I1203 16:25:10.380569 4998 generic.go:334] "Generic (PLEG): container finished" podID="c61c77db-8343-4090-ad97-06d7814f475c" containerID="bbb23e2206a12773508000e32cddaf2d381f59568d74f7a45ac7ddec5b40bf71" exitCode=0 Dec 03 16:25:10 crc kubenswrapper[4998]: I1203 16:25:10.380899 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" event={"ID":"c61c77db-8343-4090-ad97-06d7814f475c","Type":"ContainerDied","Data":"bbb23e2206a12773508000e32cddaf2d381f59568d74f7a45ac7ddec5b40bf71"} Dec 03 16:25:10 crc kubenswrapper[4998]: I1203 16:25:10.380957 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" event={"ID":"c61c77db-8343-4090-ad97-06d7814f475c","Type":"ContainerStarted","Data":"50d1c77b3705953dd177fbf55913122d319a25fa662e9b970f3c4201d941e15e"} Dec 03 16:25:10 crc kubenswrapper[4998]: I1203 16:25:10.381199 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Dec 03 16:25:10 crc kubenswrapper[4998]: I1203 16:25:10.428471 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Dec 03 16:25:10 crc kubenswrapper[4998]: I1203 16:25:10.771498 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Dec 03 16:25:10 crc kubenswrapper[4998]: I1203 16:25:10.847251 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b94f565-0948-4e4f-9636-2833d43bb829-config-data\") pod \"5b94f565-0948-4e4f-9636-2833d43bb829\" (UID: \"5b94f565-0948-4e4f-9636-2833d43bb829\") " Dec 03 16:25:10 crc kubenswrapper[4998]: I1203 16:25:10.847301 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fcmz\" (UniqueName: \"kubernetes.io/projected/5b94f565-0948-4e4f-9636-2833d43bb829-kube-api-access-8fcmz\") pod \"5b94f565-0948-4e4f-9636-2833d43bb829\" (UID: \"5b94f565-0948-4e4f-9636-2833d43bb829\") " Dec 03 16:25:10 crc kubenswrapper[4998]: I1203 16:25:10.847383 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b94f565-0948-4e4f-9636-2833d43bb829-logs\") pod \"5b94f565-0948-4e4f-9636-2833d43bb829\" (UID: \"5b94f565-0948-4e4f-9636-2833d43bb829\") " Dec 03 16:25:10 crc kubenswrapper[4998]: I1203 16:25:10.847491 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b94f565-0948-4e4f-9636-2833d43bb829-combined-ca-bundle\") pod \"5b94f565-0948-4e4f-9636-2833d43bb829\" (UID: \"5b94f565-0948-4e4f-9636-2833d43bb829\") " Dec 03 16:25:10 crc kubenswrapper[4998]: I1203 16:25:10.854329 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b94f565-0948-4e4f-9636-2833d43bb829-logs" (OuterVolumeSpecName: "logs") pod "5b94f565-0948-4e4f-9636-2833d43bb829" (UID: "5b94f565-0948-4e4f-9636-2833d43bb829"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:25:10 crc kubenswrapper[4998]: I1203 16:25:10.875126 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b94f565-0948-4e4f-9636-2833d43bb829-kube-api-access-8fcmz" (OuterVolumeSpecName: "kube-api-access-8fcmz") pod "5b94f565-0948-4e4f-9636-2833d43bb829" (UID: "5b94f565-0948-4e4f-9636-2833d43bb829"). InnerVolumeSpecName "kube-api-access-8fcmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:25:10 crc kubenswrapper[4998]: I1203 16:25:10.876824 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b94f565-0948-4e4f-9636-2833d43bb829-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b94f565-0948-4e4f-9636-2833d43bb829" (UID: "5b94f565-0948-4e4f-9636-2833d43bb829"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:10 crc kubenswrapper[4998]: I1203 16:25:10.903628 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b94f565-0948-4e4f-9636-2833d43bb829-config-data" (OuterVolumeSpecName: "config-data") pod "5b94f565-0948-4e4f-9636-2833d43bb829" (UID: "5b94f565-0948-4e4f-9636-2833d43bb829"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:10 crc kubenswrapper[4998]: I1203 16:25:10.950097 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b94f565-0948-4e4f-9636-2833d43bb829-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:10 crc kubenswrapper[4998]: I1203 16:25:10.950124 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fcmz\" (UniqueName: \"kubernetes.io/projected/5b94f565-0948-4e4f-9636-2833d43bb829-kube-api-access-8fcmz\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:10 crc kubenswrapper[4998]: I1203 16:25:10.950133 4998 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b94f565-0948-4e4f-9636-2833d43bb829-logs\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:10 crc kubenswrapper[4998]: I1203 16:25:10.950143 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b94f565-0948-4e4f-9636-2833d43bb829-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.433347 4998 generic.go:334] "Generic (PLEG): container finished" podID="e4faa24a-a43f-4d24-b20d-09e241d0cc96" containerID="a9100307a749acaf5cb3ca1c61dbd1c3014727d4fc73d649fe5f2633f22ad98e" exitCode=0 Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.433529 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4faa24a-a43f-4d24-b20d-09e241d0cc96","Type":"ContainerDied","Data":"a9100307a749acaf5cb3ca1c61dbd1c3014727d4fc73d649fe5f2633f22ad98e"} Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.436148 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"5b94f565-0948-4e4f-9636-2833d43bb829","Type":"ContainerDied","Data":"70609850792f79e03427fa3205636c4090b8ec99876d05763ae501a7d07fb7a6"} Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.436190 4998 scope.go:117] "RemoveContainer" containerID="2c4f81d00299bd0d05b7f51e1c0c848a779a1b0b6f199befff2e205150316838" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.436364 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.532002 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-applier-0"] Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.541574 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-applier-0"] Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.542569 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.558093 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-applier-0"] Dec 03 16:25:11 crc kubenswrapper[4998]: E1203 16:25:11.558530 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b94f565-0948-4e4f-9636-2833d43bb829" containerName="watcher-applier" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.558545 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b94f565-0948-4e4f-9636-2833d43bb829" containerName="watcher-applier" Dec 03 16:25:11 crc kubenswrapper[4998]: E1203 16:25:11.558555 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4faa24a-a43f-4d24-b20d-09e241d0cc96" containerName="proxy-httpd" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.558563 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4faa24a-a43f-4d24-b20d-09e241d0cc96" containerName="proxy-httpd" Dec 03 16:25:11 crc kubenswrapper[4998]: E1203 16:25:11.558589 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4faa24a-a43f-4d24-b20d-09e241d0cc96" containerName="ceilometer-central-agent" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.558599 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4faa24a-a43f-4d24-b20d-09e241d0cc96" containerName="ceilometer-central-agent" Dec 03 16:25:11 crc kubenswrapper[4998]: E1203 16:25:11.558609 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4faa24a-a43f-4d24-b20d-09e241d0cc96" containerName="sg-core" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.558633 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4faa24a-a43f-4d24-b20d-09e241d0cc96" containerName="sg-core" Dec 03 16:25:11 crc kubenswrapper[4998]: E1203 16:25:11.558660 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4faa24a-a43f-4d24-b20d-09e241d0cc96" containerName="ceilometer-notification-agent" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.558670 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4faa24a-a43f-4d24-b20d-09e241d0cc96" containerName="ceilometer-notification-agent" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.559128 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b94f565-0948-4e4f-9636-2833d43bb829" containerName="watcher-applier" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.559159 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4faa24a-a43f-4d24-b20d-09e241d0cc96" containerName="sg-core" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.559167 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4faa24a-a43f-4d24-b20d-09e241d0cc96" containerName="ceilometer-central-agent" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.559184 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4faa24a-a43f-4d24-b20d-09e241d0cc96" containerName="ceilometer-notification-agent" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.559208 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4faa24a-a43f-4d24-b20d-09e241d0cc96" containerName="proxy-httpd" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.560057 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.563794 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-applier-config-data" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.584219 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.666628 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4faa24a-a43f-4d24-b20d-09e241d0cc96-run-httpd\") pod \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.668795 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4faa24a-a43f-4d24-b20d-09e241d0cc96-config-data\") pod \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.669127 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4faa24a-a43f-4d24-b20d-09e241d0cc96-combined-ca-bundle\") pod \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.669355 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4faa24a-a43f-4d24-b20d-09e241d0cc96-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e4faa24a-a43f-4d24-b20d-09e241d0cc96" (UID: "e4faa24a-a43f-4d24-b20d-09e241d0cc96"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.669572 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4faa24a-a43f-4d24-b20d-09e241d0cc96-log-httpd\") pod \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.670572 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4faa24a-a43f-4d24-b20d-09e241d0cc96-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e4faa24a-a43f-4d24-b20d-09e241d0cc96" (UID: "e4faa24a-a43f-4d24-b20d-09e241d0cc96"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.673009 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4faa24a-a43f-4d24-b20d-09e241d0cc96-scripts\") pod \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.673542 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e4faa24a-a43f-4d24-b20d-09e241d0cc96-sg-core-conf-yaml\") pod \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.673638 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6j4p\" (UniqueName: \"kubernetes.io/projected/e4faa24a-a43f-4d24-b20d-09e241d0cc96-kube-api-access-n6j4p\") pod \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\" (UID: \"e4faa24a-a43f-4d24-b20d-09e241d0cc96\") " Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.674000 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/880abec3-6fe5-41e0-97c0-43ef4b8aedfe-config-data\") pod \"watcher-applier-0\" (UID: \"880abec3-6fe5-41e0-97c0-43ef4b8aedfe\") " pod="openstack/watcher-applier-0" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.674077 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mz9v5\" (UniqueName: \"kubernetes.io/projected/880abec3-6fe5-41e0-97c0-43ef4b8aedfe-kube-api-access-mz9v5\") pod \"watcher-applier-0\" (UID: \"880abec3-6fe5-41e0-97c0-43ef4b8aedfe\") " pod="openstack/watcher-applier-0" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.674898 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880abec3-6fe5-41e0-97c0-43ef4b8aedfe-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"880abec3-6fe5-41e0-97c0-43ef4b8aedfe\") " pod="openstack/watcher-applier-0" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.675148 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/880abec3-6fe5-41e0-97c0-43ef4b8aedfe-logs\") pod \"watcher-applier-0\" (UID: \"880abec3-6fe5-41e0-97c0-43ef4b8aedfe\") " pod="openstack/watcher-applier-0" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.675499 4998 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4faa24a-a43f-4d24-b20d-09e241d0cc96-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.675596 4998 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e4faa24a-a43f-4d24-b20d-09e241d0cc96-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.695954 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4faa24a-a43f-4d24-b20d-09e241d0cc96-scripts" (OuterVolumeSpecName: "scripts") pod "e4faa24a-a43f-4d24-b20d-09e241d0cc96" (UID: "e4faa24a-a43f-4d24-b20d-09e241d0cc96"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.697746 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4faa24a-a43f-4d24-b20d-09e241d0cc96-kube-api-access-n6j4p" (OuterVolumeSpecName: "kube-api-access-n6j4p") pod "e4faa24a-a43f-4d24-b20d-09e241d0cc96" (UID: "e4faa24a-a43f-4d24-b20d-09e241d0cc96"). InnerVolumeSpecName "kube-api-access-n6j4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.743402 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b94f565-0948-4e4f-9636-2833d43bb829" path="/var/lib/kubelet/pods/5b94f565-0948-4e4f-9636-2833d43bb829/volumes" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.786576 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/880abec3-6fe5-41e0-97c0-43ef4b8aedfe-config-data\") pod \"watcher-applier-0\" (UID: \"880abec3-6fe5-41e0-97c0-43ef4b8aedfe\") " pod="openstack/watcher-applier-0" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.787541 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mz9v5\" (UniqueName: \"kubernetes.io/projected/880abec3-6fe5-41e0-97c0-43ef4b8aedfe-kube-api-access-mz9v5\") pod \"watcher-applier-0\" (UID: \"880abec3-6fe5-41e0-97c0-43ef4b8aedfe\") " pod="openstack/watcher-applier-0" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.787642 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880abec3-6fe5-41e0-97c0-43ef4b8aedfe-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"880abec3-6fe5-41e0-97c0-43ef4b8aedfe\") " pod="openstack/watcher-applier-0" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.787807 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/880abec3-6fe5-41e0-97c0-43ef4b8aedfe-logs\") pod \"watcher-applier-0\" (UID: \"880abec3-6fe5-41e0-97c0-43ef4b8aedfe\") " pod="openstack/watcher-applier-0" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.789695 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4faa24a-a43f-4d24-b20d-09e241d0cc96-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.789922 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6j4p\" (UniqueName: \"kubernetes.io/projected/e4faa24a-a43f-4d24-b20d-09e241d0cc96-kube-api-access-n6j4p\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.788705 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/880abec3-6fe5-41e0-97c0-43ef4b8aedfe-logs\") pod \"watcher-applier-0\" (UID: \"880abec3-6fe5-41e0-97c0-43ef4b8aedfe\") " pod="openstack/watcher-applier-0" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.794912 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4faa24a-a43f-4d24-b20d-09e241d0cc96-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e4faa24a-a43f-4d24-b20d-09e241d0cc96" (UID: "e4faa24a-a43f-4d24-b20d-09e241d0cc96"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.795435 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/880abec3-6fe5-41e0-97c0-43ef4b8aedfe-config-data\") pod \"watcher-applier-0\" (UID: \"880abec3-6fe5-41e0-97c0-43ef4b8aedfe\") " pod="openstack/watcher-applier-0" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.807718 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880abec3-6fe5-41e0-97c0-43ef4b8aedfe-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"880abec3-6fe5-41e0-97c0-43ef4b8aedfe\") " pod="openstack/watcher-applier-0" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.812118 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mz9v5\" (UniqueName: \"kubernetes.io/projected/880abec3-6fe5-41e0-97c0-43ef4b8aedfe-kube-api-access-mz9v5\") pod \"watcher-applier-0\" (UID: \"880abec3-6fe5-41e0-97c0-43ef4b8aedfe\") " pod="openstack/watcher-applier-0" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.838937 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4faa24a-a43f-4d24-b20d-09e241d0cc96-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e4faa24a-a43f-4d24-b20d-09e241d0cc96" (UID: "e4faa24a-a43f-4d24-b20d-09e241d0cc96"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.860297 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4faa24a-a43f-4d24-b20d-09e241d0cc96-config-data" (OuterVolumeSpecName: "config-data") pod "e4faa24a-a43f-4d24-b20d-09e241d0cc96" (UID: "e4faa24a-a43f-4d24-b20d-09e241d0cc96"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.897385 4998 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e4faa24a-a43f-4d24-b20d-09e241d0cc96-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.897544 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4faa24a-a43f-4d24-b20d-09e241d0cc96-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.913867 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4faa24a-a43f-4d24-b20d-09e241d0cc96-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.939397 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-57f6856fd8-t44p5"] Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.941053 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-57f6856fd8-t44p5"] Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.941160 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.943259 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.943281 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 03 16:25:11 crc kubenswrapper[4998]: I1203 16:25:11.968361 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.015451 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4-logs\") pod \"barbican-api-57f6856fd8-t44p5\" (UID: \"e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4\") " pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.015568 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4-config-data\") pod \"barbican-api-57f6856fd8-t44p5\" (UID: \"e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4\") " pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.015598 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4-config-data-custom\") pod \"barbican-api-57f6856fd8-t44p5\" (UID: \"e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4\") " pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.015619 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4-combined-ca-bundle\") pod \"barbican-api-57f6856fd8-t44p5\" (UID: \"e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4\") " pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.015633 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4-public-tls-certs\") pod \"barbican-api-57f6856fd8-t44p5\" (UID: \"e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4\") " pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.015676 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsr2f\" (UniqueName: \"kubernetes.io/projected/e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4-kube-api-access-hsr2f\") pod \"barbican-api-57f6856fd8-t44p5\" (UID: \"e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4\") " pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.015694 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4-internal-tls-certs\") pod \"barbican-api-57f6856fd8-t44p5\" (UID: \"e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4\") " pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.117527 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsr2f\" (UniqueName: \"kubernetes.io/projected/e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4-kube-api-access-hsr2f\") pod \"barbican-api-57f6856fd8-t44p5\" (UID: \"e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4\") " pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.117573 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4-internal-tls-certs\") pod \"barbican-api-57f6856fd8-t44p5\" (UID: \"e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4\") " pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.117633 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4-logs\") pod \"barbican-api-57f6856fd8-t44p5\" (UID: \"e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4\") " pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.117716 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4-config-data\") pod \"barbican-api-57f6856fd8-t44p5\" (UID: \"e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4\") " pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.117742 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4-config-data-custom\") pod \"barbican-api-57f6856fd8-t44p5\" (UID: \"e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4\") " pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.117783 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4-combined-ca-bundle\") pod \"barbican-api-57f6856fd8-t44p5\" (UID: \"e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4\") " pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.117808 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4-public-tls-certs\") pod \"barbican-api-57f6856fd8-t44p5\" (UID: \"e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4\") " pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.119454 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4-logs\") pod \"barbican-api-57f6856fd8-t44p5\" (UID: \"e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4\") " pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.123865 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4-combined-ca-bundle\") pod \"barbican-api-57f6856fd8-t44p5\" (UID: \"e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4\") " pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.129373 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4-config-data\") pod \"barbican-api-57f6856fd8-t44p5\" (UID: \"e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4\") " pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.129432 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4-config-data-custom\") pod \"barbican-api-57f6856fd8-t44p5\" (UID: \"e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4\") " pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.129882 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4-public-tls-certs\") pod \"barbican-api-57f6856fd8-t44p5\" (UID: \"e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4\") " pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.130685 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4-internal-tls-certs\") pod \"barbican-api-57f6856fd8-t44p5\" (UID: \"e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4\") " pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.151438 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsr2f\" (UniqueName: \"kubernetes.io/projected/e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4-kube-api-access-hsr2f\") pod \"barbican-api-57f6856fd8-t44p5\" (UID: \"e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4\") " pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.277176 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:12 crc kubenswrapper[4998]: W1203 16:25:12.453614 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod880abec3_6fe5_41e0_97c0_43ef4b8aedfe.slice/crio-ee4dd05c843e769d1a77dd2e2a2ccf09070c5b510ac83634ee37a7c0391532f2 WatchSource:0}: Error finding container ee4dd05c843e769d1a77dd2e2a2ccf09070c5b510ac83634ee37a7c0391532f2: Status 404 returned error can't find the container with id ee4dd05c843e769d1a77dd2e2a2ccf09070c5b510ac83634ee37a7c0391532f2 Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.459212 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.479399 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" event={"ID":"c61c77db-8343-4090-ad97-06d7814f475c","Type":"ContainerStarted","Data":"76284935ee15058c4a4f41334f6b19b2e9ff45c83838ecff73ad8f612b579462"} Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.479487 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.481945 4998 generic.go:334] "Generic (PLEG): container finished" podID="a009aeee-c849-4c00-b41b-3cf5e36a75bb" containerID="616cb19e04d79255ec317f93c152a8d4bf88d542e297bf72a69bea31d5e86255" exitCode=1 Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.482006 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"a009aeee-c849-4c00-b41b-3cf5e36a75bb","Type":"ContainerDied","Data":"616cb19e04d79255ec317f93c152a8d4bf88d542e297bf72a69bea31d5e86255"} Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.482034 4998 scope.go:117] "RemoveContainer" containerID="4b34811f72882bb0f24e132377974281dacd94e6a42158df320e63964f1961b2" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.482700 4998 scope.go:117] "RemoveContainer" containerID="616cb19e04d79255ec317f93c152a8d4bf88d542e297bf72a69bea31d5e86255" Dec 03 16:25:12 crc kubenswrapper[4998]: E1203 16:25:12.482968 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 10s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(a009aeee-c849-4c00-b41b-3cf5e36a75bb)\"" pod="openstack/watcher-decision-engine-0" podUID="a009aeee-c849-4c00-b41b-3cf5e36a75bb" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.493741 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7d588c5fc8-x85fx" event={"ID":"57638b87-e1dd-4583-acc1-d57eb123de4d","Type":"ContainerStarted","Data":"7fe820a6a8b32da5c22d43fb1525fce22200190bb6a3d5a69edbd928e727dd25"} Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.493836 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7d588c5fc8-x85fx" event={"ID":"57638b87-e1dd-4583-acc1-d57eb123de4d","Type":"ContainerStarted","Data":"6eff1c0a93b954402db3e54d0f9c5acdf41f487c7f038d52b56624aa78aa00ca"} Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.500746 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" podStartSLOduration=4.500731842 podStartE2EDuration="4.500731842s" podCreationTimestamp="2025-12-03 16:25:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:25:12.499445841 +0000 UTC m=+1291.111146054" watchObservedRunningTime="2025-12-03 16:25:12.500731842 +0000 UTC m=+1291.112432065" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.502795 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6768bc7bbd-hsnsm" event={"ID":"fda7bb47-a754-4123-b031-14030f822167","Type":"ContainerStarted","Data":"3661a2f2fc4d9f66e967e4d98638f1e6d4087ebe0e74f4e46ce5d3c3f02c542c"} Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.503922 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6768bc7bbd-hsnsm" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.503972 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6768bc7bbd-hsnsm" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.508598 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7d6d5bf649-h8gnx" event={"ID":"d33fe6f6-a02c-4841-baf4-1c956cd2e9f2","Type":"ContainerStarted","Data":"276f0f2c3b3f34ede349d6829ddb0d06e7391ea1b468589a17f525d47bbedbc2"} Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.508642 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7d6d5bf649-h8gnx" event={"ID":"d33fe6f6-a02c-4841-baf4-1c956cd2e9f2","Type":"ContainerStarted","Data":"855c2e3a9059478cac6ea86f0504768d34b411b3e53cdbfadd69b22dd4e8f799"} Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.521836 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e4faa24a-a43f-4d24-b20d-09e241d0cc96","Type":"ContainerDied","Data":"0475afca48189c206af41d74b245bf5418653f1ca935fd76768cd39cef6f38b0"} Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.521966 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.542265 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7d588c5fc8-x85fx" podStartSLOduration=2.81840721 podStartE2EDuration="4.542240742s" podCreationTimestamp="2025-12-03 16:25:08 +0000 UTC" firstStartedPulling="2025-12-03 16:25:09.593419931 +0000 UTC m=+1288.205120144" lastFinishedPulling="2025-12-03 16:25:11.317253453 +0000 UTC m=+1289.928953676" observedRunningTime="2025-12-03 16:25:12.537112987 +0000 UTC m=+1291.148813210" watchObservedRunningTime="2025-12-03 16:25:12.542240742 +0000 UTC m=+1291.153940965" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.606706 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6768bc7bbd-hsnsm" podStartSLOduration=4.606685209 podStartE2EDuration="4.606685209s" podCreationTimestamp="2025-12-03 16:25:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:25:12.595635971 +0000 UTC m=+1291.207336194" watchObservedRunningTime="2025-12-03 16:25:12.606685209 +0000 UTC m=+1291.218385432" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.610222 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7d6d5bf649-h8gnx" podStartSLOduration=2.886765302 podStartE2EDuration="4.610207655s" podCreationTimestamp="2025-12-03 16:25:08 +0000 UTC" firstStartedPulling="2025-12-03 16:25:09.592784025 +0000 UTC m=+1288.204484248" lastFinishedPulling="2025-12-03 16:25:11.316226378 +0000 UTC m=+1289.927926601" observedRunningTime="2025-12-03 16:25:12.579952279 +0000 UTC m=+1291.191652502" watchObservedRunningTime="2025-12-03 16:25:12.610207655 +0000 UTC m=+1291.221907878" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.709599 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.727214 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.748937 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.755015 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.765467 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.766216 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.780046 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.788043 4998 scope.go:117] "RemoveContainer" containerID="f4e16b7cd25c4006759bea8a0ecae336123d5c61dcd9a285eafc5e9515aa15d4" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.794855 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-57f6856fd8-t44p5"] Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.852285 4998 scope.go:117] "RemoveContainer" containerID="6f0d6533d024330f91e4fcc8a4ad9f21d7ed7fb31327d7ca16b483a8f88d0308" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.856676 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/855663b4-d4de-4ca0-811c-44d4def02706-log-httpd\") pod \"ceilometer-0\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " pod="openstack/ceilometer-0" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.856836 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/855663b4-d4de-4ca0-811c-44d4def02706-scripts\") pod \"ceilometer-0\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " pod="openstack/ceilometer-0" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.856951 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/855663b4-d4de-4ca0-811c-44d4def02706-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " pod="openstack/ceilometer-0" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.857073 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/855663b4-d4de-4ca0-811c-44d4def02706-config-data\") pod \"ceilometer-0\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " pod="openstack/ceilometer-0" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.857201 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/855663b4-d4de-4ca0-811c-44d4def02706-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " pod="openstack/ceilometer-0" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.857365 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/855663b4-d4de-4ca0-811c-44d4def02706-run-httpd\") pod \"ceilometer-0\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " pod="openstack/ceilometer-0" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.857469 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czzv5\" (UniqueName: \"kubernetes.io/projected/855663b4-d4de-4ca0-811c-44d4def02706-kube-api-access-czzv5\") pod \"ceilometer-0\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " pod="openstack/ceilometer-0" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.885843 4998 scope.go:117] "RemoveContainer" containerID="a9100307a749acaf5cb3ca1c61dbd1c3014727d4fc73d649fe5f2633f22ad98e" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.951988 4998 scope.go:117] "RemoveContainer" containerID="5ae86f8014172d9b95a4494ff98b6cbc3dfc4a7399092fa9f3e71c9260a39e54" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.958706 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/855663b4-d4de-4ca0-811c-44d4def02706-run-httpd\") pod \"ceilometer-0\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " pod="openstack/ceilometer-0" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.958737 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czzv5\" (UniqueName: \"kubernetes.io/projected/855663b4-d4de-4ca0-811c-44d4def02706-kube-api-access-czzv5\") pod \"ceilometer-0\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " pod="openstack/ceilometer-0" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.958803 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/855663b4-d4de-4ca0-811c-44d4def02706-log-httpd\") pod \"ceilometer-0\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " pod="openstack/ceilometer-0" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.958834 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/855663b4-d4de-4ca0-811c-44d4def02706-scripts\") pod \"ceilometer-0\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " pod="openstack/ceilometer-0" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.958859 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/855663b4-d4de-4ca0-811c-44d4def02706-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " pod="openstack/ceilometer-0" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.958903 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/855663b4-d4de-4ca0-811c-44d4def02706-config-data\") pod \"ceilometer-0\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " pod="openstack/ceilometer-0" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.958938 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/855663b4-d4de-4ca0-811c-44d4def02706-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " pod="openstack/ceilometer-0" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.959470 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/855663b4-d4de-4ca0-811c-44d4def02706-log-httpd\") pod \"ceilometer-0\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " pod="openstack/ceilometer-0" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.959725 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/855663b4-d4de-4ca0-811c-44d4def02706-run-httpd\") pod \"ceilometer-0\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " pod="openstack/ceilometer-0" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.968906 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/855663b4-d4de-4ca0-811c-44d4def02706-config-data\") pod \"ceilometer-0\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " pod="openstack/ceilometer-0" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.969138 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/855663b4-d4de-4ca0-811c-44d4def02706-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " pod="openstack/ceilometer-0" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.969668 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/855663b4-d4de-4ca0-811c-44d4def02706-scripts\") pod \"ceilometer-0\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " pod="openstack/ceilometer-0" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.981826 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czzv5\" (UniqueName: \"kubernetes.io/projected/855663b4-d4de-4ca0-811c-44d4def02706-kube-api-access-czzv5\") pod \"ceilometer-0\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " pod="openstack/ceilometer-0" Dec 03 16:25:12 crc kubenswrapper[4998]: I1203 16:25:12.984930 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/855663b4-d4de-4ca0-811c-44d4def02706-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " pod="openstack/ceilometer-0" Dec 03 16:25:13 crc kubenswrapper[4998]: I1203 16:25:13.076787 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:25:13 crc kubenswrapper[4998]: I1203 16:25:13.531648 4998 scope.go:117] "RemoveContainer" containerID="616cb19e04d79255ec317f93c152a8d4bf88d542e297bf72a69bea31d5e86255" Dec 03 16:25:13 crc kubenswrapper[4998]: E1203 16:25:13.532164 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 10s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(a009aeee-c849-4c00-b41b-3cf5e36a75bb)\"" pod="openstack/watcher-decision-engine-0" podUID="a009aeee-c849-4c00-b41b-3cf5e36a75bb" Dec 03 16:25:13 crc kubenswrapper[4998]: I1203 16:25:13.535034 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57f6856fd8-t44p5" event={"ID":"e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4","Type":"ContainerStarted","Data":"8f5ad38074d9e519736d194b6ad6ad6c12386f62e7e664c677b9c34f1c266e6b"} Dec 03 16:25:13 crc kubenswrapper[4998]: I1203 16:25:13.538019 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"880abec3-6fe5-41e0-97c0-43ef4b8aedfe","Type":"ContainerStarted","Data":"16ef45f54f9c9847b35b13c66b3b0296de2bb1007301b6577e28f5ec15a23e4b"} Dec 03 16:25:13 crc kubenswrapper[4998]: I1203 16:25:13.538061 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"880abec3-6fe5-41e0-97c0-43ef4b8aedfe","Type":"ContainerStarted","Data":"ee4dd05c843e769d1a77dd2e2a2ccf09070c5b510ac83634ee37a7c0391532f2"} Dec 03 16:25:13 crc kubenswrapper[4998]: I1203 16:25:13.604156 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:25:13 crc kubenswrapper[4998]: W1203 16:25:13.610947 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod855663b4_d4de_4ca0_811c_44d4def02706.slice/crio-3b776860dd5eb21a0151994321cc0d35a66bb4babdf3376145b3fd6ebef47320 WatchSource:0}: Error finding container 3b776860dd5eb21a0151994321cc0d35a66bb4babdf3376145b3fd6ebef47320: Status 404 returned error can't find the container with id 3b776860dd5eb21a0151994321cc0d35a66bb4babdf3376145b3fd6ebef47320 Dec 03 16:25:13 crc kubenswrapper[4998]: I1203 16:25:13.692804 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4faa24a-a43f-4d24-b20d-09e241d0cc96" path="/var/lib/kubelet/pods/e4faa24a-a43f-4d24-b20d-09e241d0cc96/volumes" Dec 03 16:25:14 crc kubenswrapper[4998]: I1203 16:25:14.549048 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"855663b4-d4de-4ca0-811c-44d4def02706","Type":"ContainerStarted","Data":"3b776860dd5eb21a0151994321cc0d35a66bb4babdf3376145b3fd6ebef47320"} Dec 03 16:25:15 crc kubenswrapper[4998]: I1203 16:25:15.565260 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57f6856fd8-t44p5" event={"ID":"e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4","Type":"ContainerStarted","Data":"260d9511259ba3e8e8ecb3ccd7fafcff32edca34987b3f6b3a83f5a24bf380a9"} Dec 03 16:25:15 crc kubenswrapper[4998]: I1203 16:25:15.598442 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-applier-0" podStartSLOduration=4.598413924 podStartE2EDuration="4.598413924s" podCreationTimestamp="2025-12-03 16:25:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:25:15.587928709 +0000 UTC m=+1294.199628932" watchObservedRunningTime="2025-12-03 16:25:15.598413924 +0000 UTC m=+1294.210114177" Dec 03 16:25:16 crc kubenswrapper[4998]: I1203 16:25:16.574811 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"855663b4-d4de-4ca0-811c-44d4def02706","Type":"ContainerStarted","Data":"2c63eb0b5ccdeff1d9234d2c6cc21f2aab83a23b017acd58cc9a301f5ca959c6"} Dec 03 16:25:16 crc kubenswrapper[4998]: I1203 16:25:16.575530 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"855663b4-d4de-4ca0-811c-44d4def02706","Type":"ContainerStarted","Data":"bd642294c5a767432620ab6245505b9509dcd42d5a1e0c7143d68ee5baa5b382"} Dec 03 16:25:16 crc kubenswrapper[4998]: I1203 16:25:16.577105 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57f6856fd8-t44p5" event={"ID":"e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4","Type":"ContainerStarted","Data":"0976cf3808ee5e247d53053e89971f5239e420bf03083410bfc935948081133b"} Dec 03 16:25:16 crc kubenswrapper[4998]: I1203 16:25:16.577404 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:16 crc kubenswrapper[4998]: I1203 16:25:16.577432 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:16 crc kubenswrapper[4998]: I1203 16:25:16.597883 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7565d78f8d-jm9qt" podUID="d7600639-7175-4c0a-9524-40adb505d3ea" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.159:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.159:8443: connect: connection refused" Dec 03 16:25:16 crc kubenswrapper[4998]: I1203 16:25:16.610027 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-57f6856fd8-t44p5" podStartSLOduration=5.610003982 podStartE2EDuration="5.610003982s" podCreationTimestamp="2025-12-03 16:25:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:25:16.592112777 +0000 UTC m=+1295.203813010" watchObservedRunningTime="2025-12-03 16:25:16.610003982 +0000 UTC m=+1295.221704205" Dec 03 16:25:16 crc kubenswrapper[4998]: I1203 16:25:16.969447 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-applier-0" Dec 03 16:25:17 crc kubenswrapper[4998]: I1203 16:25:17.592012 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"855663b4-d4de-4ca0-811c-44d4def02706","Type":"ContainerStarted","Data":"d467c7642ab7b67ec0bb6e94bf7954b5748a94ef142fb733a8fe87b353b6838d"} Dec 03 16:25:17 crc kubenswrapper[4998]: I1203 16:25:17.596403 4998 generic.go:334] "Generic (PLEG): container finished" podID="82aa9d23-457c-45a0-9729-e5a803738237" containerID="2298bd80e5bdaea4315dbb4b6a0effa05cbdcae8e440ead4f9287b4f7bd6ed75" exitCode=0 Dec 03 16:25:17 crc kubenswrapper[4998]: I1203 16:25:17.596873 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8lz9j" event={"ID":"82aa9d23-457c-45a0-9729-e5a803738237","Type":"ContainerDied","Data":"2298bd80e5bdaea4315dbb4b6a0effa05cbdcae8e440ead4f9287b4f7bd6ed75"} Dec 03 16:25:18 crc kubenswrapper[4998]: I1203 16:25:18.609033 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"855663b4-d4de-4ca0-811c-44d4def02706","Type":"ContainerStarted","Data":"f75e25cdac20077210bd49dc9952da9b52dc8acf1d5253701e494daa9d79e244"} Dec 03 16:25:18 crc kubenswrapper[4998]: I1203 16:25:18.609369 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 16:25:18 crc kubenswrapper[4998]: I1203 16:25:18.644060 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.977839503 podStartE2EDuration="6.64403946s" podCreationTimestamp="2025-12-03 16:25:12 +0000 UTC" firstStartedPulling="2025-12-03 16:25:13.613659364 +0000 UTC m=+1292.225359627" lastFinishedPulling="2025-12-03 16:25:18.279859351 +0000 UTC m=+1296.891559584" observedRunningTime="2025-12-03 16:25:18.635546454 +0000 UTC m=+1297.247246677" watchObservedRunningTime="2025-12-03 16:25:18.64403946 +0000 UTC m=+1297.255739683" Dec 03 16:25:18 crc kubenswrapper[4998]: I1203 16:25:18.995993 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8lz9j" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.087898 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnlxq\" (UniqueName: \"kubernetes.io/projected/82aa9d23-457c-45a0-9729-e5a803738237-kube-api-access-jnlxq\") pod \"82aa9d23-457c-45a0-9729-e5a803738237\" (UID: \"82aa9d23-457c-45a0-9729-e5a803738237\") " Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.088013 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82aa9d23-457c-45a0-9729-e5a803738237-config-data\") pod \"82aa9d23-457c-45a0-9729-e5a803738237\" (UID: \"82aa9d23-457c-45a0-9729-e5a803738237\") " Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.088108 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82aa9d23-457c-45a0-9729-e5a803738237-scripts\") pod \"82aa9d23-457c-45a0-9729-e5a803738237\" (UID: \"82aa9d23-457c-45a0-9729-e5a803738237\") " Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.088147 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/82aa9d23-457c-45a0-9729-e5a803738237-db-sync-config-data\") pod \"82aa9d23-457c-45a0-9729-e5a803738237\" (UID: \"82aa9d23-457c-45a0-9729-e5a803738237\") " Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.088224 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82aa9d23-457c-45a0-9729-e5a803738237-combined-ca-bundle\") pod \"82aa9d23-457c-45a0-9729-e5a803738237\" (UID: \"82aa9d23-457c-45a0-9729-e5a803738237\") " Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.088244 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82aa9d23-457c-45a0-9729-e5a803738237-etc-machine-id\") pod \"82aa9d23-457c-45a0-9729-e5a803738237\" (UID: \"82aa9d23-457c-45a0-9729-e5a803738237\") " Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.088606 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/82aa9d23-457c-45a0-9729-e5a803738237-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "82aa9d23-457c-45a0-9729-e5a803738237" (UID: "82aa9d23-457c-45a0-9729-e5a803738237"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.093301 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82aa9d23-457c-45a0-9729-e5a803738237-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "82aa9d23-457c-45a0-9729-e5a803738237" (UID: "82aa9d23-457c-45a0-9729-e5a803738237"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.093940 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82aa9d23-457c-45a0-9729-e5a803738237-kube-api-access-jnlxq" (OuterVolumeSpecName: "kube-api-access-jnlxq") pod "82aa9d23-457c-45a0-9729-e5a803738237" (UID: "82aa9d23-457c-45a0-9729-e5a803738237"). InnerVolumeSpecName "kube-api-access-jnlxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.095877 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82aa9d23-457c-45a0-9729-e5a803738237-scripts" (OuterVolumeSpecName: "scripts") pod "82aa9d23-457c-45a0-9729-e5a803738237" (UID: "82aa9d23-457c-45a0-9729-e5a803738237"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.137353 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82aa9d23-457c-45a0-9729-e5a803738237-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82aa9d23-457c-45a0-9729-e5a803738237" (UID: "82aa9d23-457c-45a0-9729-e5a803738237"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.153847 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82aa9d23-457c-45a0-9729-e5a803738237-config-data" (OuterVolumeSpecName: "config-data") pod "82aa9d23-457c-45a0-9729-e5a803738237" (UID: "82aa9d23-457c-45a0-9729-e5a803738237"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.189820 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82aa9d23-457c-45a0-9729-e5a803738237-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.189849 4998 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/82aa9d23-457c-45a0-9729-e5a803738237-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.189860 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82aa9d23-457c-45a0-9729-e5a803738237-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.189869 4998 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82aa9d23-457c-45a0-9729-e5a803738237-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.189878 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnlxq\" (UniqueName: \"kubernetes.io/projected/82aa9d23-457c-45a0-9729-e5a803738237-kube-api-access-jnlxq\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.189886 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82aa9d23-457c-45a0-9729-e5a803738237-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.243046 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.351468 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dd959b98c-6ffpz"] Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.351847 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" podUID="6f1c1a9e-4b7e-474a-99c9-0b79b259f774" containerName="dnsmasq-dns" containerID="cri-o://e4dd17f34f33175e96a5ce0731b4953eafffeff8eb564be760614bdde21c1e9c" gracePeriod=10 Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.517888 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.518947 4998 scope.go:117] "RemoveContainer" containerID="616cb19e04d79255ec317f93c152a8d4bf88d542e297bf72a69bea31d5e86255" Dec 03 16:25:19 crc kubenswrapper[4998]: E1203 16:25:19.519226 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 10s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(a009aeee-c849-4c00-b41b-3cf5e36a75bb)\"" pod="openstack/watcher-decision-engine-0" podUID="a009aeee-c849-4c00-b41b-3cf5e36a75bb" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.619995 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8lz9j" event={"ID":"82aa9d23-457c-45a0-9729-e5a803738237","Type":"ContainerDied","Data":"f5c3e814f3475f33ec9ee3b208a68df32ef8b689c79d08eff56c919ab7090191"} Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.620036 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5c3e814f3475f33ec9ee3b208a68df32ef8b689c79d08eff56c919ab7090191" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.620010 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8lz9j" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.621797 4998 generic.go:334] "Generic (PLEG): container finished" podID="6f1c1a9e-4b7e-474a-99c9-0b79b259f774" containerID="e4dd17f34f33175e96a5ce0731b4953eafffeff8eb564be760614bdde21c1e9c" exitCode=0 Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.622386 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" event={"ID":"6f1c1a9e-4b7e-474a-99c9-0b79b259f774","Type":"ContainerDied","Data":"e4dd17f34f33175e96a5ce0731b4953eafffeff8eb564be760614bdde21c1e9c"} Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.922499 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7874d7c47f-wsqpb"] Dec 03 16:25:19 crc kubenswrapper[4998]: E1203 16:25:19.923041 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82aa9d23-457c-45a0-9729-e5a803738237" containerName="cinder-db-sync" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.923060 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="82aa9d23-457c-45a0-9729-e5a803738237" containerName="cinder-db-sync" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.923227 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="82aa9d23-457c-45a0-9729-e5a803738237" containerName="cinder-db-sync" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.924237 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.932455 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.935054 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.946600 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-8dr6s" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.946871 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.946989 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.947137 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.986977 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" Dec 03 16:25:19 crc kubenswrapper[4998]: I1203 16:25:19.999831 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.043650 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vjxd\" (UniqueName: \"kubernetes.io/projected/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-kube-api-access-7vjxd\") pod \"dnsmasq-dns-7874d7c47f-wsqpb\" (UID: \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\") " pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.044066 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-ovsdbserver-nb\") pod \"dnsmasq-dns-7874d7c47f-wsqpb\" (UID: \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\") " pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.044143 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-dns-svc\") pod \"dnsmasq-dns-7874d7c47f-wsqpb\" (UID: \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\") " pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.044170 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-ovsdbserver-sb\") pod \"dnsmasq-dns-7874d7c47f-wsqpb\" (UID: \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\") " pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.044206 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-config\") pod \"dnsmasq-dns-7874d7c47f-wsqpb\" (UID: \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\") " pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.044273 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-dns-swift-storage-0\") pod \"dnsmasq-dns-7874d7c47f-wsqpb\" (UID: \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\") " pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.044312 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6872e78-e3b5-4024-9de0-24c6a3cfe345-config-data\") pod \"cinder-scheduler-0\" (UID: \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.044347 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e6872e78-e3b5-4024-9de0-24c6a3cfe345-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.044375 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e6872e78-e3b5-4024-9de0-24c6a3cfe345-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.044399 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6872e78-e3b5-4024-9de0-24c6a3cfe345-scripts\") pod \"cinder-scheduler-0\" (UID: \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.044456 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6872e78-e3b5-4024-9de0-24c6a3cfe345-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.044509 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66ggp\" (UniqueName: \"kubernetes.io/projected/e6872e78-e3b5-4024-9de0-24c6a3cfe345-kube-api-access-66ggp\") pod \"cinder-scheduler-0\" (UID: \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.093294 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7874d7c47f-wsqpb"] Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.134300 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 16:25:20 crc kubenswrapper[4998]: E1203 16:25:20.140023 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f1c1a9e-4b7e-474a-99c9-0b79b259f774" containerName="dnsmasq-dns" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.140057 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f1c1a9e-4b7e-474a-99c9-0b79b259f774" containerName="dnsmasq-dns" Dec 03 16:25:20 crc kubenswrapper[4998]: E1203 16:25:20.140102 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f1c1a9e-4b7e-474a-99c9-0b79b259f774" containerName="init" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.140109 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f1c1a9e-4b7e-474a-99c9-0b79b259f774" containerName="init" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.140467 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f1c1a9e-4b7e-474a-99c9-0b79b259f774" containerName="dnsmasq-dns" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.141605 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.145512 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-ovsdbserver-sb\") pod \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\" (UID: \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\") " Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.145671 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gcsz\" (UniqueName: \"kubernetes.io/projected/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-kube-api-access-9gcsz\") pod \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\" (UID: \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\") " Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.145697 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-ovsdbserver-nb\") pod \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\" (UID: \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\") " Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.145758 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-dns-swift-storage-0\") pod \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\" (UID: \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\") " Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.145819 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-dns-svc\") pod \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\" (UID: \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\") " Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.145853 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-config\") pod \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\" (UID: \"6f1c1a9e-4b7e-474a-99c9-0b79b259f774\") " Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.146080 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66ggp\" (UniqueName: \"kubernetes.io/projected/e6872e78-e3b5-4024-9de0-24c6a3cfe345-kube-api-access-66ggp\") pod \"cinder-scheduler-0\" (UID: \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.146137 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vjxd\" (UniqueName: \"kubernetes.io/projected/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-kube-api-access-7vjxd\") pod \"dnsmasq-dns-7874d7c47f-wsqpb\" (UID: \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\") " pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.146173 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-ovsdbserver-nb\") pod \"dnsmasq-dns-7874d7c47f-wsqpb\" (UID: \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\") " pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.146215 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-dns-svc\") pod \"dnsmasq-dns-7874d7c47f-wsqpb\" (UID: \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\") " pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.146217 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.146230 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-ovsdbserver-sb\") pod \"dnsmasq-dns-7874d7c47f-wsqpb\" (UID: \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\") " pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.146981 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-config\") pod \"dnsmasq-dns-7874d7c47f-wsqpb\" (UID: \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\") " pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.147099 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-dns-swift-storage-0\") pod \"dnsmasq-dns-7874d7c47f-wsqpb\" (UID: \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\") " pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.147121 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-ovsdbserver-sb\") pod \"dnsmasq-dns-7874d7c47f-wsqpb\" (UID: \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\") " pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.147159 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6872e78-e3b5-4024-9de0-24c6a3cfe345-config-data\") pod \"cinder-scheduler-0\" (UID: \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.147211 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e6872e78-e3b5-4024-9de0-24c6a3cfe345-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.147244 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e6872e78-e3b5-4024-9de0-24c6a3cfe345-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.147274 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6872e78-e3b5-4024-9de0-24c6a3cfe345-scripts\") pod \"cinder-scheduler-0\" (UID: \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.147450 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6872e78-e3b5-4024-9de0-24c6a3cfe345-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.153978 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-ovsdbserver-nb\") pod \"dnsmasq-dns-7874d7c47f-wsqpb\" (UID: \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\") " pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.156449 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-dns-svc\") pod \"dnsmasq-dns-7874d7c47f-wsqpb\" (UID: \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\") " pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.158743 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6872e78-e3b5-4024-9de0-24c6a3cfe345-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.159386 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-config\") pod \"dnsmasq-dns-7874d7c47f-wsqpb\" (UID: \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\") " pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.159930 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-dns-swift-storage-0\") pod \"dnsmasq-dns-7874d7c47f-wsqpb\" (UID: \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\") " pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.161212 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e6872e78-e3b5-4024-9de0-24c6a3cfe345-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.163413 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-kube-api-access-9gcsz" (OuterVolumeSpecName: "kube-api-access-9gcsz") pod "6f1c1a9e-4b7e-474a-99c9-0b79b259f774" (UID: "6f1c1a9e-4b7e-474a-99c9-0b79b259f774"). InnerVolumeSpecName "kube-api-access-9gcsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.178740 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6872e78-e3b5-4024-9de0-24c6a3cfe345-config-data\") pod \"cinder-scheduler-0\" (UID: \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.183262 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e6872e78-e3b5-4024-9de0-24c6a3cfe345-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.198312 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6872e78-e3b5-4024-9de0-24c6a3cfe345-scripts\") pod \"cinder-scheduler-0\" (UID: \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.208987 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.210049 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vjxd\" (UniqueName: \"kubernetes.io/projected/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-kube-api-access-7vjxd\") pod \"dnsmasq-dns-7874d7c47f-wsqpb\" (UID: \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\") " pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.230325 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66ggp\" (UniqueName: \"kubernetes.io/projected/e6872e78-e3b5-4024-9de0-24c6a3cfe345-kube-api-access-66ggp\") pod \"cinder-scheduler-0\" (UID: \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.249366 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-config-data\") pod \"cinder-api-0\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " pod="openstack/cinder-api-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.249421 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-config-data-custom\") pod \"cinder-api-0\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " pod="openstack/cinder-api-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.249449 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " pod="openstack/cinder-api-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.249476 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " pod="openstack/cinder-api-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.249492 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-logs\") pod \"cinder-api-0\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " pod="openstack/cinder-api-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.249514 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7gkh\" (UniqueName: \"kubernetes.io/projected/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-kube-api-access-v7gkh\") pod \"cinder-api-0\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " pod="openstack/cinder-api-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.249566 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-scripts\") pod \"cinder-api-0\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " pod="openstack/cinder-api-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.249715 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gcsz\" (UniqueName: \"kubernetes.io/projected/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-kube-api-access-9gcsz\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.283261 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6f1c1a9e-4b7e-474a-99c9-0b79b259f774" (UID: "6f1c1a9e-4b7e-474a-99c9-0b79b259f774"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.310908 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-config" (OuterVolumeSpecName: "config") pod "6f1c1a9e-4b7e-474a-99c9-0b79b259f774" (UID: "6f1c1a9e-4b7e-474a-99c9-0b79b259f774"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.346452 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6f1c1a9e-4b7e-474a-99c9-0b79b259f774" (UID: "6f1c1a9e-4b7e-474a-99c9-0b79b259f774"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.349163 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.353368 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-config-data\") pod \"cinder-api-0\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " pod="openstack/cinder-api-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.353416 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-config-data-custom\") pod \"cinder-api-0\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " pod="openstack/cinder-api-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.353441 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " pod="openstack/cinder-api-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.353465 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " pod="openstack/cinder-api-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.353480 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-logs\") pod \"cinder-api-0\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " pod="openstack/cinder-api-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.353499 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7gkh\" (UniqueName: \"kubernetes.io/projected/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-kube-api-access-v7gkh\") pod \"cinder-api-0\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " pod="openstack/cinder-api-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.353534 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-scripts\") pod \"cinder-api-0\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " pod="openstack/cinder-api-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.353586 4998 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.353597 4998 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.353606 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.356785 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " pod="openstack/cinder-api-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.360607 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-config-data\") pod \"cinder-api-0\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " pod="openstack/cinder-api-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.362179 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-logs\") pod \"cinder-api-0\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " pod="openstack/cinder-api-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.363313 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.364749 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-scripts\") pod \"cinder-api-0\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " pod="openstack/cinder-api-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.373450 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " pod="openstack/cinder-api-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.404118 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-config-data-custom\") pod \"cinder-api-0\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " pod="openstack/cinder-api-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.414873 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6f1c1a9e-4b7e-474a-99c9-0b79b259f774" (UID: "6f1c1a9e-4b7e-474a-99c9-0b79b259f774"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.415243 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6f1c1a9e-4b7e-474a-99c9-0b79b259f774" (UID: "6f1c1a9e-4b7e-474a-99c9-0b79b259f774"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.432344 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7gkh\" (UniqueName: \"kubernetes.io/projected/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-kube-api-access-v7gkh\") pod \"cinder-api-0\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " pod="openstack/cinder-api-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.455987 4998 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.456028 4998 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6f1c1a9e-4b7e-474a-99c9-0b79b259f774-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.511531 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.639855 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" event={"ID":"6f1c1a9e-4b7e-474a-99c9-0b79b259f774","Type":"ContainerDied","Data":"986b94e956c674309235ca750ba3b6e7573e938fe15d44013cb5fb07b9c49c76"} Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.640206 4998 scope.go:117] "RemoveContainer" containerID="e4dd17f34f33175e96a5ce0731b4953eafffeff8eb564be760614bdde21c1e9c" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.640386 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dd959b98c-6ffpz" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.690720 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dd959b98c-6ffpz"] Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.710570 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-dd959b98c-6ffpz"] Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.715195 4998 scope.go:117] "RemoveContainer" containerID="e10ef0292f2e0c2ec6385af6e9ca41e1eb4a1620a7c75659d92488e45404d91c" Dec 03 16:25:20 crc kubenswrapper[4998]: I1203 16:25:20.981861 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7874d7c47f-wsqpb"] Dec 03 16:25:21 crc kubenswrapper[4998]: I1203 16:25:21.029446 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6764c5c65f-jtlz7" Dec 03 16:25:21 crc kubenswrapper[4998]: I1203 16:25:21.185453 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 16:25:21 crc kubenswrapper[4998]: I1203 16:25:21.210846 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 16:25:21 crc kubenswrapper[4998]: I1203 16:25:21.652091 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf","Type":"ContainerStarted","Data":"e4e63573804de55c264f465658b61f086c9b0740e45b8c509f6f0e81a143e022"} Dec 03 16:25:21 crc kubenswrapper[4998]: I1203 16:25:21.654090 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e6872e78-e3b5-4024-9de0-24c6a3cfe345","Type":"ContainerStarted","Data":"6825d6228b457e2e86c5bf832da1451c4e304e994cb5de0693806ec8f3ad36d4"} Dec 03 16:25:21 crc kubenswrapper[4998]: I1203 16:25:21.655408 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" event={"ID":"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae","Type":"ContainerStarted","Data":"42d9340e89ad7aa1fb42965aa482053d3962c701861edba6fada3ef0cf8c4288"} Dec 03 16:25:21 crc kubenswrapper[4998]: I1203 16:25:21.767538 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f1c1a9e-4b7e-474a-99c9-0b79b259f774" path="/var/lib/kubelet/pods/6f1c1a9e-4b7e-474a-99c9-0b79b259f774/volumes" Dec 03 16:25:21 crc kubenswrapper[4998]: I1203 16:25:21.811222 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6768bc7bbd-hsnsm" Dec 03 16:25:21 crc kubenswrapper[4998]: I1203 16:25:21.969195 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-applier-0" Dec 03 16:25:22 crc kubenswrapper[4998]: I1203 16:25:22.016393 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-applier-0" Dec 03 16:25:22 crc kubenswrapper[4998]: I1203 16:25:22.592814 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6768bc7bbd-hsnsm" Dec 03 16:25:22 crc kubenswrapper[4998]: I1203 16:25:22.712481 4998 generic.go:334] "Generic (PLEG): container finished" podID="cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae" containerID="c7e9d13c1e7428c03ce2084b599648401e4458737c45ced94dce153ccce8e09a" exitCode=0 Dec 03 16:25:22 crc kubenswrapper[4998]: I1203 16:25:22.713528 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" event={"ID":"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae","Type":"ContainerDied","Data":"c7e9d13c1e7428c03ce2084b599648401e4458737c45ced94dce153ccce8e09a"} Dec 03 16:25:22 crc kubenswrapper[4998]: I1203 16:25:22.867004 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-applier-0" Dec 03 16:25:23 crc kubenswrapper[4998]: I1203 16:25:23.077551 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 16:25:23 crc kubenswrapper[4998]: I1203 16:25:23.729040 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf","Type":"ContainerStarted","Data":"4cd5621bd9999f8912e9f8dac7be546d02d473e54439ae066cb6d6d394507220"} Dec 03 16:25:24 crc kubenswrapper[4998]: I1203 16:25:24.744496 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf","Type":"ContainerStarted","Data":"a02f38a8cb552441e6687ec55547b52c4b41a36f038a3ee855fa5659d41077a3"} Dec 03 16:25:24 crc kubenswrapper[4998]: I1203 16:25:24.745087 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 16:25:24 crc kubenswrapper[4998]: I1203 16:25:24.744813 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf" containerName="cinder-api" containerID="cri-o://a02f38a8cb552441e6687ec55547b52c4b41a36f038a3ee855fa5659d41077a3" gracePeriod=30 Dec 03 16:25:24 crc kubenswrapper[4998]: I1203 16:25:24.744589 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf" containerName="cinder-api-log" containerID="cri-o://4cd5621bd9999f8912e9f8dac7be546d02d473e54439ae066cb6d6d394507220" gracePeriod=30 Dec 03 16:25:24 crc kubenswrapper[4998]: I1203 16:25:24.755872 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e6872e78-e3b5-4024-9de0-24c6a3cfe345","Type":"ContainerStarted","Data":"9a133c288c85b51f5f5a60e0fe7913b20f9fd4a9bf8104561e23e33479dada46"} Dec 03 16:25:24 crc kubenswrapper[4998]: I1203 16:25:24.755931 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e6872e78-e3b5-4024-9de0-24c6a3cfe345","Type":"ContainerStarted","Data":"47ad5ff218d1db4e4a839220fd027f05fd704ae47e711f9827767480718bfd25"} Dec 03 16:25:24 crc kubenswrapper[4998]: I1203 16:25:24.758405 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" event={"ID":"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae","Type":"ContainerStarted","Data":"975d5e8e1934c341117b10d8de86a23b2e1de3acee36deaac7f11e29bd7c5895"} Dec 03 16:25:24 crc kubenswrapper[4998]: I1203 16:25:24.758845 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" Dec 03 16:25:24 crc kubenswrapper[4998]: I1203 16:25:24.765616 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.76560581 podStartE2EDuration="4.76560581s" podCreationTimestamp="2025-12-03 16:25:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:25:24.760777582 +0000 UTC m=+1303.372477815" watchObservedRunningTime="2025-12-03 16:25:24.76560581 +0000 UTC m=+1303.377306033" Dec 03 16:25:24 crc kubenswrapper[4998]: I1203 16:25:24.782737 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.733150174 podStartE2EDuration="5.782719876s" podCreationTimestamp="2025-12-03 16:25:19 +0000 UTC" firstStartedPulling="2025-12-03 16:25:21.186002004 +0000 UTC m=+1299.797702227" lastFinishedPulling="2025-12-03 16:25:22.235571706 +0000 UTC m=+1300.847271929" observedRunningTime="2025-12-03 16:25:24.777604982 +0000 UTC m=+1303.389305215" watchObservedRunningTime="2025-12-03 16:25:24.782719876 +0000 UTC m=+1303.394420089" Dec 03 16:25:24 crc kubenswrapper[4998]: I1203 16:25:24.804932 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" podStartSLOduration=5.804913156 podStartE2EDuration="5.804913156s" podCreationTimestamp="2025-12-03 16:25:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:25:24.798176162 +0000 UTC m=+1303.409876385" watchObservedRunningTime="2025-12-03 16:25:24.804913156 +0000 UTC m=+1303.416613379" Dec 03 16:25:24 crc kubenswrapper[4998]: I1203 16:25:24.905553 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:25 crc kubenswrapper[4998]: I1203 16:25:25.250913 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-57f6856fd8-t44p5" Dec 03 16:25:25 crc kubenswrapper[4998]: I1203 16:25:25.323921 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6768bc7bbd-hsnsm"] Dec 03 16:25:25 crc kubenswrapper[4998]: I1203 16:25:25.324393 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6768bc7bbd-hsnsm" podUID="fda7bb47-a754-4123-b031-14030f822167" containerName="barbican-api-log" containerID="cri-o://9d3f8d796e12715831aedb33ba672326901694d1c6e8c35f8ebcf992ac8abcdd" gracePeriod=30 Dec 03 16:25:25 crc kubenswrapper[4998]: I1203 16:25:25.324601 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6768bc7bbd-hsnsm" podUID="fda7bb47-a754-4123-b031-14030f822167" containerName="barbican-api" containerID="cri-o://3661a2f2fc4d9f66e967e4d98638f1e6d4087ebe0e74f4e46ce5d3c3f02c542c" gracePeriod=30 Dec 03 16:25:25 crc kubenswrapper[4998]: I1203 16:25:25.365238 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 16:25:25 crc kubenswrapper[4998]: I1203 16:25:25.757792 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 03 16:25:25 crc kubenswrapper[4998]: I1203 16:25:25.759210 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 16:25:25 crc kubenswrapper[4998]: I1203 16:25:25.761110 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 03 16:25:25 crc kubenswrapper[4998]: I1203 16:25:25.761284 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-s62q7" Dec 03 16:25:25 crc kubenswrapper[4998]: I1203 16:25:25.761435 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 03 16:25:25 crc kubenswrapper[4998]: I1203 16:25:25.785857 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 16:25:25 crc kubenswrapper[4998]: I1203 16:25:25.811291 4998 generic.go:334] "Generic (PLEG): container finished" podID="a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf" containerID="4cd5621bd9999f8912e9f8dac7be546d02d473e54439ae066cb6d6d394507220" exitCode=143 Dec 03 16:25:25 crc kubenswrapper[4998]: I1203 16:25:25.811379 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf","Type":"ContainerDied","Data":"4cd5621bd9999f8912e9f8dac7be546d02d473e54439ae066cb6d6d394507220"} Dec 03 16:25:25 crc kubenswrapper[4998]: I1203 16:25:25.825349 4998 generic.go:334] "Generic (PLEG): container finished" podID="ac00cda5-18ab-48de-a82e-1ebe08fa518c" containerID="d9de792d4b5c1b7f62defd70fa80c00e1a269c701e996345d575192a2ef656f2" exitCode=0 Dec 03 16:25:25 crc kubenswrapper[4998]: I1203 16:25:25.825608 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2g9pt" event={"ID":"ac00cda5-18ab-48de-a82e-1ebe08fa518c","Type":"ContainerDied","Data":"d9de792d4b5c1b7f62defd70fa80c00e1a269c701e996345d575192a2ef656f2"} Dec 03 16:25:25 crc kubenswrapper[4998]: I1203 16:25:25.837540 4998 generic.go:334] "Generic (PLEG): container finished" podID="fda7bb47-a754-4123-b031-14030f822167" containerID="9d3f8d796e12715831aedb33ba672326901694d1c6e8c35f8ebcf992ac8abcdd" exitCode=143 Dec 03 16:25:25 crc kubenswrapper[4998]: I1203 16:25:25.837605 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6768bc7bbd-hsnsm" event={"ID":"fda7bb47-a754-4123-b031-14030f822167","Type":"ContainerDied","Data":"9d3f8d796e12715831aedb33ba672326901694d1c6e8c35f8ebcf992ac8abcdd"} Dec 03 16:25:25 crc kubenswrapper[4998]: I1203 16:25:25.933375 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/041471bf-1760-4fb8-ac9a-81dc1dd7defa-openstack-config\") pod \"openstackclient\" (UID: \"041471bf-1760-4fb8-ac9a-81dc1dd7defa\") " pod="openstack/openstackclient" Dec 03 16:25:25 crc kubenswrapper[4998]: I1203 16:25:25.933453 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhnp9\" (UniqueName: \"kubernetes.io/projected/041471bf-1760-4fb8-ac9a-81dc1dd7defa-kube-api-access-dhnp9\") pod \"openstackclient\" (UID: \"041471bf-1760-4fb8-ac9a-81dc1dd7defa\") " pod="openstack/openstackclient" Dec 03 16:25:25 crc kubenswrapper[4998]: I1203 16:25:25.933644 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/041471bf-1760-4fb8-ac9a-81dc1dd7defa-openstack-config-secret\") pod \"openstackclient\" (UID: \"041471bf-1760-4fb8-ac9a-81dc1dd7defa\") " pod="openstack/openstackclient" Dec 03 16:25:25 crc kubenswrapper[4998]: I1203 16:25:25.933745 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/041471bf-1760-4fb8-ac9a-81dc1dd7defa-combined-ca-bundle\") pod \"openstackclient\" (UID: \"041471bf-1760-4fb8-ac9a-81dc1dd7defa\") " pod="openstack/openstackclient" Dec 03 16:25:26 crc kubenswrapper[4998]: I1203 16:25:26.034898 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/041471bf-1760-4fb8-ac9a-81dc1dd7defa-openstack-config-secret\") pod \"openstackclient\" (UID: \"041471bf-1760-4fb8-ac9a-81dc1dd7defa\") " pod="openstack/openstackclient" Dec 03 16:25:26 crc kubenswrapper[4998]: I1203 16:25:26.035006 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/041471bf-1760-4fb8-ac9a-81dc1dd7defa-combined-ca-bundle\") pod \"openstackclient\" (UID: \"041471bf-1760-4fb8-ac9a-81dc1dd7defa\") " pod="openstack/openstackclient" Dec 03 16:25:26 crc kubenswrapper[4998]: I1203 16:25:26.035045 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/041471bf-1760-4fb8-ac9a-81dc1dd7defa-openstack-config\") pod \"openstackclient\" (UID: \"041471bf-1760-4fb8-ac9a-81dc1dd7defa\") " pod="openstack/openstackclient" Dec 03 16:25:26 crc kubenswrapper[4998]: I1203 16:25:26.035068 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhnp9\" (UniqueName: \"kubernetes.io/projected/041471bf-1760-4fb8-ac9a-81dc1dd7defa-kube-api-access-dhnp9\") pod \"openstackclient\" (UID: \"041471bf-1760-4fb8-ac9a-81dc1dd7defa\") " pod="openstack/openstackclient" Dec 03 16:25:26 crc kubenswrapper[4998]: I1203 16:25:26.037249 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/041471bf-1760-4fb8-ac9a-81dc1dd7defa-openstack-config\") pod \"openstackclient\" (UID: \"041471bf-1760-4fb8-ac9a-81dc1dd7defa\") " pod="openstack/openstackclient" Dec 03 16:25:26 crc kubenswrapper[4998]: I1203 16:25:26.040572 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/041471bf-1760-4fb8-ac9a-81dc1dd7defa-combined-ca-bundle\") pod \"openstackclient\" (UID: \"041471bf-1760-4fb8-ac9a-81dc1dd7defa\") " pod="openstack/openstackclient" Dec 03 16:25:26 crc kubenswrapper[4998]: I1203 16:25:26.040897 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/041471bf-1760-4fb8-ac9a-81dc1dd7defa-openstack-config-secret\") pod \"openstackclient\" (UID: \"041471bf-1760-4fb8-ac9a-81dc1dd7defa\") " pod="openstack/openstackclient" Dec 03 16:25:26 crc kubenswrapper[4998]: I1203 16:25:26.056370 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhnp9\" (UniqueName: \"kubernetes.io/projected/041471bf-1760-4fb8-ac9a-81dc1dd7defa-kube-api-access-dhnp9\") pod \"openstackclient\" (UID: \"041471bf-1760-4fb8-ac9a-81dc1dd7defa\") " pod="openstack/openstackclient" Dec 03 16:25:26 crc kubenswrapper[4998]: I1203 16:25:26.097281 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 16:25:26 crc kubenswrapper[4998]: I1203 16:25:26.598366 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7565d78f8d-jm9qt" podUID="d7600639-7175-4c0a-9524-40adb505d3ea" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.159:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.159:8443: connect: connection refused" Dec 03 16:25:26 crc kubenswrapper[4998]: I1203 16:25:26.599158 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:25:26 crc kubenswrapper[4998]: I1203 16:25:26.731955 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 16:25:26 crc kubenswrapper[4998]: I1203 16:25:26.854863 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"041471bf-1760-4fb8-ac9a-81dc1dd7defa","Type":"ContainerStarted","Data":"96931ca14a23ee1929ee67b0602bba2603b800bfb165ce3f1f5555ceb66b1b9c"} Dec 03 16:25:27 crc kubenswrapper[4998]: I1203 16:25:27.113244 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:25:27 crc kubenswrapper[4998]: I1203 16:25:27.113557 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:25:27 crc kubenswrapper[4998]: I1203 16:25:27.113599 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:25:27 crc kubenswrapper[4998]: I1203 16:25:27.114330 4998 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dc47f739351bd606edc721ca357f0c89b466b48ca6131ea4bbb968b4e4cc02c4"} pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 16:25:27 crc kubenswrapper[4998]: I1203 16:25:27.114403 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" containerID="cri-o://dc47f739351bd606edc721ca357f0c89b466b48ca6131ea4bbb968b4e4cc02c4" gracePeriod=600 Dec 03 16:25:27 crc kubenswrapper[4998]: I1203 16:25:27.289625 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2g9pt" Dec 03 16:25:27 crc kubenswrapper[4998]: I1203 16:25:27.468967 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggth6\" (UniqueName: \"kubernetes.io/projected/ac00cda5-18ab-48de-a82e-1ebe08fa518c-kube-api-access-ggth6\") pod \"ac00cda5-18ab-48de-a82e-1ebe08fa518c\" (UID: \"ac00cda5-18ab-48de-a82e-1ebe08fa518c\") " Dec 03 16:25:27 crc kubenswrapper[4998]: I1203 16:25:27.469155 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ac00cda5-18ab-48de-a82e-1ebe08fa518c-config\") pod \"ac00cda5-18ab-48de-a82e-1ebe08fa518c\" (UID: \"ac00cda5-18ab-48de-a82e-1ebe08fa518c\") " Dec 03 16:25:27 crc kubenswrapper[4998]: I1203 16:25:27.469250 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac00cda5-18ab-48de-a82e-1ebe08fa518c-combined-ca-bundle\") pod \"ac00cda5-18ab-48de-a82e-1ebe08fa518c\" (UID: \"ac00cda5-18ab-48de-a82e-1ebe08fa518c\") " Dec 03 16:25:27 crc kubenswrapper[4998]: I1203 16:25:27.494616 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac00cda5-18ab-48de-a82e-1ebe08fa518c-kube-api-access-ggth6" (OuterVolumeSpecName: "kube-api-access-ggth6") pod "ac00cda5-18ab-48de-a82e-1ebe08fa518c" (UID: "ac00cda5-18ab-48de-a82e-1ebe08fa518c"). InnerVolumeSpecName "kube-api-access-ggth6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:25:27 crc kubenswrapper[4998]: I1203 16:25:27.511999 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac00cda5-18ab-48de-a82e-1ebe08fa518c-config" (OuterVolumeSpecName: "config") pod "ac00cda5-18ab-48de-a82e-1ebe08fa518c" (UID: "ac00cda5-18ab-48de-a82e-1ebe08fa518c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:27 crc kubenswrapper[4998]: I1203 16:25:27.515413 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac00cda5-18ab-48de-a82e-1ebe08fa518c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac00cda5-18ab-48de-a82e-1ebe08fa518c" (UID: "ac00cda5-18ab-48de-a82e-1ebe08fa518c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:27 crc kubenswrapper[4998]: I1203 16:25:27.571973 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac00cda5-18ab-48de-a82e-1ebe08fa518c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:27 crc kubenswrapper[4998]: I1203 16:25:27.571998 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggth6\" (UniqueName: \"kubernetes.io/projected/ac00cda5-18ab-48de-a82e-1ebe08fa518c-kube-api-access-ggth6\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:27 crc kubenswrapper[4998]: I1203 16:25:27.572007 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/ac00cda5-18ab-48de-a82e-1ebe08fa518c-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:27 crc kubenswrapper[4998]: I1203 16:25:27.885362 4998 generic.go:334] "Generic (PLEG): container finished" podID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerID="dc47f739351bd606edc721ca357f0c89b466b48ca6131ea4bbb968b4e4cc02c4" exitCode=0 Dec 03 16:25:27 crc kubenswrapper[4998]: I1203 16:25:27.887070 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerDied","Data":"dc47f739351bd606edc721ca357f0c89b466b48ca6131ea4bbb968b4e4cc02c4"} Dec 03 16:25:27 crc kubenswrapper[4998]: I1203 16:25:27.887275 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379"} Dec 03 16:25:27 crc kubenswrapper[4998]: I1203 16:25:27.887351 4998 scope.go:117] "RemoveContainer" containerID="4e819484c40f765e06b9e1c5a211c516f5099460ab695ea890ca1412f47f0680" Dec 03 16:25:27 crc kubenswrapper[4998]: I1203 16:25:27.897649 4998 generic.go:334] "Generic (PLEG): container finished" podID="de77cff6-7039-418c-8114-3c64ea4ce8e4" containerID="043af9beddc6495b4dd97ae2a362ea5bea94ef2a3d8e62d695b5f5765c933eda" exitCode=0 Dec 03 16:25:27 crc kubenswrapper[4998]: I1203 16:25:27.897998 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-2rqhm" event={"ID":"de77cff6-7039-418c-8114-3c64ea4ce8e4","Type":"ContainerDied","Data":"043af9beddc6495b4dd97ae2a362ea5bea94ef2a3d8e62d695b5f5765c933eda"} Dec 03 16:25:27 crc kubenswrapper[4998]: I1203 16:25:27.916143 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2g9pt" event={"ID":"ac00cda5-18ab-48de-a82e-1ebe08fa518c","Type":"ContainerDied","Data":"24f750ec22c49315b4d52610f475929dea6e8d7ec23668f695b057b4182a54ef"} Dec 03 16:25:27 crc kubenswrapper[4998]: I1203 16:25:27.916445 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24f750ec22c49315b4d52610f475929dea6e8d7ec23668f695b057b4182a54ef" Dec 03 16:25:27 crc kubenswrapper[4998]: I1203 16:25:27.916620 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2g9pt" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.128072 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7874d7c47f-wsqpb"] Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.128291 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" podUID="cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae" containerName="dnsmasq-dns" containerID="cri-o://975d5e8e1934c341117b10d8de86a23b2e1de3acee36deaac7f11e29bd7c5895" gracePeriod=10 Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.132131 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.170234 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55646d875-2lpr8"] Dec 03 16:25:28 crc kubenswrapper[4998]: E1203 16:25:28.170646 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac00cda5-18ab-48de-a82e-1ebe08fa518c" containerName="neutron-db-sync" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.170658 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac00cda5-18ab-48de-a82e-1ebe08fa518c" containerName="neutron-db-sync" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.170853 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac00cda5-18ab-48de-a82e-1ebe08fa518c" containerName="neutron-db-sync" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.173782 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55646d875-2lpr8" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.213401 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7f47995564-5mql2"] Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.215094 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7f47995564-5mql2" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.219684 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.219933 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.220107 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-f986n" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.226679 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.289736 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-ovsdbserver-nb\") pod \"dnsmasq-dns-55646d875-2lpr8\" (UID: \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\") " pod="openstack/dnsmasq-dns-55646d875-2lpr8" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.290108 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-ovsdbserver-sb\") pod \"dnsmasq-dns-55646d875-2lpr8\" (UID: \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\") " pod="openstack/dnsmasq-dns-55646d875-2lpr8" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.290298 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-config\") pod \"dnsmasq-dns-55646d875-2lpr8\" (UID: \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\") " pod="openstack/dnsmasq-dns-55646d875-2lpr8" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.290061 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55646d875-2lpr8"] Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.290409 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-dns-svc\") pod \"dnsmasq-dns-55646d875-2lpr8\" (UID: \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\") " pod="openstack/dnsmasq-dns-55646d875-2lpr8" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.290541 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfqrv\" (UniqueName: \"kubernetes.io/projected/72f65fee-d135-44e7-92c5-4f1afb25c2ab-kube-api-access-mfqrv\") pod \"dnsmasq-dns-55646d875-2lpr8\" (UID: \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\") " pod="openstack/dnsmasq-dns-55646d875-2lpr8" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.290585 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-dns-swift-storage-0\") pod \"dnsmasq-dns-55646d875-2lpr8\" (UID: \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\") " pod="openstack/dnsmasq-dns-55646d875-2lpr8" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.305699 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7f47995564-5mql2"] Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.392357 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98bb0f8f-aa52-4408-8d16-c6677b080101-combined-ca-bundle\") pod \"neutron-7f47995564-5mql2\" (UID: \"98bb0f8f-aa52-4408-8d16-c6677b080101\") " pod="openstack/neutron-7f47995564-5mql2" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.392419 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/98bb0f8f-aa52-4408-8d16-c6677b080101-config\") pod \"neutron-7f47995564-5mql2\" (UID: \"98bb0f8f-aa52-4408-8d16-c6677b080101\") " pod="openstack/neutron-7f47995564-5mql2" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.392451 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/98bb0f8f-aa52-4408-8d16-c6677b080101-ovndb-tls-certs\") pod \"neutron-7f47995564-5mql2\" (UID: \"98bb0f8f-aa52-4408-8d16-c6677b080101\") " pod="openstack/neutron-7f47995564-5mql2" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.392506 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-ovsdbserver-nb\") pod \"dnsmasq-dns-55646d875-2lpr8\" (UID: \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\") " pod="openstack/dnsmasq-dns-55646d875-2lpr8" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.392544 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/98bb0f8f-aa52-4408-8d16-c6677b080101-httpd-config\") pod \"neutron-7f47995564-5mql2\" (UID: \"98bb0f8f-aa52-4408-8d16-c6677b080101\") " pod="openstack/neutron-7f47995564-5mql2" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.392576 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-ovsdbserver-sb\") pod \"dnsmasq-dns-55646d875-2lpr8\" (UID: \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\") " pod="openstack/dnsmasq-dns-55646d875-2lpr8" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.392626 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsmw6\" (UniqueName: \"kubernetes.io/projected/98bb0f8f-aa52-4408-8d16-c6677b080101-kube-api-access-tsmw6\") pod \"neutron-7f47995564-5mql2\" (UID: \"98bb0f8f-aa52-4408-8d16-c6677b080101\") " pod="openstack/neutron-7f47995564-5mql2" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.392656 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-config\") pod \"dnsmasq-dns-55646d875-2lpr8\" (UID: \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\") " pod="openstack/dnsmasq-dns-55646d875-2lpr8" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.392682 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-dns-svc\") pod \"dnsmasq-dns-55646d875-2lpr8\" (UID: \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\") " pod="openstack/dnsmasq-dns-55646d875-2lpr8" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.392708 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfqrv\" (UniqueName: \"kubernetes.io/projected/72f65fee-d135-44e7-92c5-4f1afb25c2ab-kube-api-access-mfqrv\") pod \"dnsmasq-dns-55646d875-2lpr8\" (UID: \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\") " pod="openstack/dnsmasq-dns-55646d875-2lpr8" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.392728 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-dns-swift-storage-0\") pod \"dnsmasq-dns-55646d875-2lpr8\" (UID: \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\") " pod="openstack/dnsmasq-dns-55646d875-2lpr8" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.394089 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-dns-swift-storage-0\") pod \"dnsmasq-dns-55646d875-2lpr8\" (UID: \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\") " pod="openstack/dnsmasq-dns-55646d875-2lpr8" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.398400 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-config\") pod \"dnsmasq-dns-55646d875-2lpr8\" (UID: \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\") " pod="openstack/dnsmasq-dns-55646d875-2lpr8" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.399117 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-dns-svc\") pod \"dnsmasq-dns-55646d875-2lpr8\" (UID: \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\") " pod="openstack/dnsmasq-dns-55646d875-2lpr8" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.399205 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-ovsdbserver-nb\") pod \"dnsmasq-dns-55646d875-2lpr8\" (UID: \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\") " pod="openstack/dnsmasq-dns-55646d875-2lpr8" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.401056 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-ovsdbserver-sb\") pod \"dnsmasq-dns-55646d875-2lpr8\" (UID: \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\") " pod="openstack/dnsmasq-dns-55646d875-2lpr8" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.425660 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfqrv\" (UniqueName: \"kubernetes.io/projected/72f65fee-d135-44e7-92c5-4f1afb25c2ab-kube-api-access-mfqrv\") pod \"dnsmasq-dns-55646d875-2lpr8\" (UID: \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\") " pod="openstack/dnsmasq-dns-55646d875-2lpr8" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.494569 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/98bb0f8f-aa52-4408-8d16-c6677b080101-httpd-config\") pod \"neutron-7f47995564-5mql2\" (UID: \"98bb0f8f-aa52-4408-8d16-c6677b080101\") " pod="openstack/neutron-7f47995564-5mql2" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.495459 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsmw6\" (UniqueName: \"kubernetes.io/projected/98bb0f8f-aa52-4408-8d16-c6677b080101-kube-api-access-tsmw6\") pod \"neutron-7f47995564-5mql2\" (UID: \"98bb0f8f-aa52-4408-8d16-c6677b080101\") " pod="openstack/neutron-7f47995564-5mql2" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.495576 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98bb0f8f-aa52-4408-8d16-c6677b080101-combined-ca-bundle\") pod \"neutron-7f47995564-5mql2\" (UID: \"98bb0f8f-aa52-4408-8d16-c6677b080101\") " pod="openstack/neutron-7f47995564-5mql2" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.495620 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/98bb0f8f-aa52-4408-8d16-c6677b080101-config\") pod \"neutron-7f47995564-5mql2\" (UID: \"98bb0f8f-aa52-4408-8d16-c6677b080101\") " pod="openstack/neutron-7f47995564-5mql2" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.495639 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/98bb0f8f-aa52-4408-8d16-c6677b080101-ovndb-tls-certs\") pod \"neutron-7f47995564-5mql2\" (UID: \"98bb0f8f-aa52-4408-8d16-c6677b080101\") " pod="openstack/neutron-7f47995564-5mql2" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.499967 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98bb0f8f-aa52-4408-8d16-c6677b080101-combined-ca-bundle\") pod \"neutron-7f47995564-5mql2\" (UID: \"98bb0f8f-aa52-4408-8d16-c6677b080101\") " pod="openstack/neutron-7f47995564-5mql2" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.500329 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/98bb0f8f-aa52-4408-8d16-c6677b080101-httpd-config\") pod \"neutron-7f47995564-5mql2\" (UID: \"98bb0f8f-aa52-4408-8d16-c6677b080101\") " pod="openstack/neutron-7f47995564-5mql2" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.500830 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/98bb0f8f-aa52-4408-8d16-c6677b080101-config\") pod \"neutron-7f47995564-5mql2\" (UID: \"98bb0f8f-aa52-4408-8d16-c6677b080101\") " pod="openstack/neutron-7f47995564-5mql2" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.510513 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/98bb0f8f-aa52-4408-8d16-c6677b080101-ovndb-tls-certs\") pod \"neutron-7f47995564-5mql2\" (UID: \"98bb0f8f-aa52-4408-8d16-c6677b080101\") " pod="openstack/neutron-7f47995564-5mql2" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.517457 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsmw6\" (UniqueName: \"kubernetes.io/projected/98bb0f8f-aa52-4408-8d16-c6677b080101-kube-api-access-tsmw6\") pod \"neutron-7f47995564-5mql2\" (UID: \"98bb0f8f-aa52-4408-8d16-c6677b080101\") " pod="openstack/neutron-7f47995564-5mql2" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.641076 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55646d875-2lpr8" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.661064 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7f47995564-5mql2" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.741036 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.911588 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-ovsdbserver-nb\") pod \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\" (UID: \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\") " Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.911870 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-dns-swift-storage-0\") pod \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\" (UID: \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\") " Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.911905 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-config\") pod \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\" (UID: \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\") " Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.911921 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-dns-svc\") pod \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\" (UID: \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\") " Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.911981 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vjxd\" (UniqueName: \"kubernetes.io/projected/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-kube-api-access-7vjxd\") pod \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\" (UID: \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\") " Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.912017 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-ovsdbserver-sb\") pod \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\" (UID: \"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae\") " Dec 03 16:25:28 crc kubenswrapper[4998]: I1203 16:25:28.950050 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-kube-api-access-7vjxd" (OuterVolumeSpecName: "kube-api-access-7vjxd") pod "cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae" (UID: "cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae"). InnerVolumeSpecName "kube-api-access-7vjxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.012185 4998 generic.go:334] "Generic (PLEG): container finished" podID="cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae" containerID="975d5e8e1934c341117b10d8de86a23b2e1de3acee36deaac7f11e29bd7c5895" exitCode=0 Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.012264 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.012283 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" event={"ID":"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae","Type":"ContainerDied","Data":"975d5e8e1934c341117b10d8de86a23b2e1de3acee36deaac7f11e29bd7c5895"} Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.018815 4998 scope.go:117] "RemoveContainer" containerID="975d5e8e1934c341117b10d8de86a23b2e1de3acee36deaac7f11e29bd7c5895" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.020777 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vjxd\" (UniqueName: \"kubernetes.io/projected/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-kube-api-access-7vjxd\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.024125 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7874d7c47f-wsqpb" event={"ID":"cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae","Type":"ContainerDied","Data":"42d9340e89ad7aa1fb42965aa482053d3962c701861edba6fada3ef0cf8c4288"} Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.056302 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae" (UID: "cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.095209 4998 scope.go:117] "RemoveContainer" containerID="c7e9d13c1e7428c03ce2084b599648401e4458737c45ced94dce153ccce8e09a" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.112282 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae" (UID: "cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.118304 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae" (UID: "cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.123610 4998 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.123639 4998 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.123648 4998 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.169047 4998 scope.go:117] "RemoveContainer" containerID="975d5e8e1934c341117b10d8de86a23b2e1de3acee36deaac7f11e29bd7c5895" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.170123 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae" (UID: "cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:29 crc kubenswrapper[4998]: E1203 16:25:29.180407 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"975d5e8e1934c341117b10d8de86a23b2e1de3acee36deaac7f11e29bd7c5895\": container with ID starting with 975d5e8e1934c341117b10d8de86a23b2e1de3acee36deaac7f11e29bd7c5895 not found: ID does not exist" containerID="975d5e8e1934c341117b10d8de86a23b2e1de3acee36deaac7f11e29bd7c5895" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.180462 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"975d5e8e1934c341117b10d8de86a23b2e1de3acee36deaac7f11e29bd7c5895"} err="failed to get container status \"975d5e8e1934c341117b10d8de86a23b2e1de3acee36deaac7f11e29bd7c5895\": rpc error: code = NotFound desc = could not find container \"975d5e8e1934c341117b10d8de86a23b2e1de3acee36deaac7f11e29bd7c5895\": container with ID starting with 975d5e8e1934c341117b10d8de86a23b2e1de3acee36deaac7f11e29bd7c5895 not found: ID does not exist" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.180494 4998 scope.go:117] "RemoveContainer" containerID="c7e9d13c1e7428c03ce2084b599648401e4458737c45ced94dce153ccce8e09a" Dec 03 16:25:29 crc kubenswrapper[4998]: E1203 16:25:29.182972 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7e9d13c1e7428c03ce2084b599648401e4458737c45ced94dce153ccce8e09a\": container with ID starting with c7e9d13c1e7428c03ce2084b599648401e4458737c45ced94dce153ccce8e09a not found: ID does not exist" containerID="c7e9d13c1e7428c03ce2084b599648401e4458737c45ced94dce153ccce8e09a" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.183017 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7e9d13c1e7428c03ce2084b599648401e4458737c45ced94dce153ccce8e09a"} err="failed to get container status \"c7e9d13c1e7428c03ce2084b599648401e4458737c45ced94dce153ccce8e09a\": rpc error: code = NotFound desc = could not find container \"c7e9d13c1e7428c03ce2084b599648401e4458737c45ced94dce153ccce8e09a\": container with ID starting with c7e9d13c1e7428c03ce2084b599648401e4458737c45ced94dce153ccce8e09a not found: ID does not exist" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.203102 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-config" (OuterVolumeSpecName: "config") pod "cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae" (UID: "cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.225625 4998 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.225662 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.260336 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6768bc7bbd-hsnsm" podUID="fda7bb47-a754-4123-b031-14030f822167" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.170:9311/healthcheck\": dial tcp 10.217.0.170:9311: connect: connection refused" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.260400 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6768bc7bbd-hsnsm" podUID="fda7bb47-a754-4123-b031-14030f822167" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.170:9311/healthcheck\": dial tcp 10.217.0.170:9311: connect: connection refused" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.265887 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55646d875-2lpr8"] Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.468984 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7874d7c47f-wsqpb"] Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.507440 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7874d7c47f-wsqpb"] Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.517630 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.517679 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/watcher-decision-engine-0" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.520818 4998 scope.go:117] "RemoveContainer" containerID="616cb19e04d79255ec317f93c152a8d4bf88d542e297bf72a69bea31d5e86255" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.643265 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7f47995564-5mql2"] Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.707507 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae" path="/var/lib/kubelet/pods/cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae/volumes" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.749248 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-2rqhm" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.857330 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/de77cff6-7039-418c-8114-3c64ea4ce8e4-db-sync-config-data\") pod \"de77cff6-7039-418c-8114-3c64ea4ce8e4\" (UID: \"de77cff6-7039-418c-8114-3c64ea4ce8e4\") " Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.857621 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de77cff6-7039-418c-8114-3c64ea4ce8e4-config-data\") pod \"de77cff6-7039-418c-8114-3c64ea4ce8e4\" (UID: \"de77cff6-7039-418c-8114-3c64ea4ce8e4\") " Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.857807 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de77cff6-7039-418c-8114-3c64ea4ce8e4-combined-ca-bundle\") pod \"de77cff6-7039-418c-8114-3c64ea4ce8e4\" (UID: \"de77cff6-7039-418c-8114-3c64ea4ce8e4\") " Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.857858 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftjd7\" (UniqueName: \"kubernetes.io/projected/de77cff6-7039-418c-8114-3c64ea4ce8e4-kube-api-access-ftjd7\") pod \"de77cff6-7039-418c-8114-3c64ea4ce8e4\" (UID: \"de77cff6-7039-418c-8114-3c64ea4ce8e4\") " Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.869393 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de77cff6-7039-418c-8114-3c64ea4ce8e4-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "de77cff6-7039-418c-8114-3c64ea4ce8e4" (UID: "de77cff6-7039-418c-8114-3c64ea4ce8e4"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.874482 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de77cff6-7039-418c-8114-3c64ea4ce8e4-kube-api-access-ftjd7" (OuterVolumeSpecName: "kube-api-access-ftjd7") pod "de77cff6-7039-418c-8114-3c64ea4ce8e4" (UID: "de77cff6-7039-418c-8114-3c64ea4ce8e4"). InnerVolumeSpecName "kube-api-access-ftjd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.891703 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6768bc7bbd-hsnsm" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.916997 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de77cff6-7039-418c-8114-3c64ea4ce8e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de77cff6-7039-418c-8114-3c64ea4ce8e4" (UID: "de77cff6-7039-418c-8114-3c64ea4ce8e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.961730 4998 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/de77cff6-7039-418c-8114-3c64ea4ce8e4-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.961775 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de77cff6-7039-418c-8114-3c64ea4ce8e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.961787 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftjd7\" (UniqueName: \"kubernetes.io/projected/de77cff6-7039-418c-8114-3c64ea4ce8e4-kube-api-access-ftjd7\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:29 crc kubenswrapper[4998]: I1203 16:25:29.973331 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de77cff6-7039-418c-8114-3c64ea4ce8e4-config-data" (OuterVolumeSpecName: "config-data") pod "de77cff6-7039-418c-8114-3c64ea4ce8e4" (UID: "de77cff6-7039-418c-8114-3c64ea4ce8e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.063364 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fda7bb47-a754-4123-b031-14030f822167-config-data\") pod \"fda7bb47-a754-4123-b031-14030f822167\" (UID: \"fda7bb47-a754-4123-b031-14030f822167\") " Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.063541 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fda7bb47-a754-4123-b031-14030f822167-combined-ca-bundle\") pod \"fda7bb47-a754-4123-b031-14030f822167\" (UID: \"fda7bb47-a754-4123-b031-14030f822167\") " Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.063588 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fda7bb47-a754-4123-b031-14030f822167-config-data-custom\") pod \"fda7bb47-a754-4123-b031-14030f822167\" (UID: \"fda7bb47-a754-4123-b031-14030f822167\") " Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.063623 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fda7bb47-a754-4123-b031-14030f822167-logs\") pod \"fda7bb47-a754-4123-b031-14030f822167\" (UID: \"fda7bb47-a754-4123-b031-14030f822167\") " Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.063814 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q69g9\" (UniqueName: \"kubernetes.io/projected/fda7bb47-a754-4123-b031-14030f822167-kube-api-access-q69g9\") pod \"fda7bb47-a754-4123-b031-14030f822167\" (UID: \"fda7bb47-a754-4123-b031-14030f822167\") " Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.064285 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de77cff6-7039-418c-8114-3c64ea4ce8e4-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.065500 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fda7bb47-a754-4123-b031-14030f822167-logs" (OuterVolumeSpecName: "logs") pod "fda7bb47-a754-4123-b031-14030f822167" (UID: "fda7bb47-a754-4123-b031-14030f822167"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.070042 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda7bb47-a754-4123-b031-14030f822167-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fda7bb47-a754-4123-b031-14030f822167" (UID: "fda7bb47-a754-4123-b031-14030f822167"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.073834 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda7bb47-a754-4123-b031-14030f822167-kube-api-access-q69g9" (OuterVolumeSpecName: "kube-api-access-q69g9") pod "fda7bb47-a754-4123-b031-14030f822167" (UID: "fda7bb47-a754-4123-b031-14030f822167"). InnerVolumeSpecName "kube-api-access-q69g9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.102420 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f47995564-5mql2" event={"ID":"98bb0f8f-aa52-4408-8d16-c6677b080101","Type":"ContainerStarted","Data":"5ab8347b81ea9f11541eee83ec6cd743c0c89c981fcef844dbf98fb32b6c3599"} Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.102457 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f47995564-5mql2" event={"ID":"98bb0f8f-aa52-4408-8d16-c6677b080101","Type":"ContainerStarted","Data":"c458dd983344d1d0d82c52ca5b5caabfef781971f092857283c28e64df5fbe35"} Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.117018 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-2rqhm" event={"ID":"de77cff6-7039-418c-8114-3c64ea4ce8e4","Type":"ContainerDied","Data":"49b4dd73ac30d196a03ffee484d4181b02cb8c15f2de3aceb1023b7abb6709a4"} Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.117075 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49b4dd73ac30d196a03ffee484d4181b02cb8c15f2de3aceb1023b7abb6709a4" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.117037 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-2rqhm" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.120260 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"a009aeee-c849-4c00-b41b-3cf5e36a75bb","Type":"ContainerStarted","Data":"b657ed22e9096f43b7242cbed330ab7de4b3e3c8d87141fb5d8555f0560c2457"} Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.122651 4998 generic.go:334] "Generic (PLEG): container finished" podID="72f65fee-d135-44e7-92c5-4f1afb25c2ab" containerID="b4142ae26f57f5d78cff31bfad2f6cadffe6d46285dfcc3318fe104337bd5dc5" exitCode=0 Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.122691 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55646d875-2lpr8" event={"ID":"72f65fee-d135-44e7-92c5-4f1afb25c2ab","Type":"ContainerDied","Data":"b4142ae26f57f5d78cff31bfad2f6cadffe6d46285dfcc3318fe104337bd5dc5"} Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.122709 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55646d875-2lpr8" event={"ID":"72f65fee-d135-44e7-92c5-4f1afb25c2ab","Type":"ContainerStarted","Data":"e200a60c2efa9938c91ad44f5024e971c63a2c15fbb13bcb36c942aba3869614"} Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.125162 4998 generic.go:334] "Generic (PLEG): container finished" podID="fda7bb47-a754-4123-b031-14030f822167" containerID="3661a2f2fc4d9f66e967e4d98638f1e6d4087ebe0e74f4e46ce5d3c3f02c542c" exitCode=0 Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.125199 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6768bc7bbd-hsnsm" event={"ID":"fda7bb47-a754-4123-b031-14030f822167","Type":"ContainerDied","Data":"3661a2f2fc4d9f66e967e4d98638f1e6d4087ebe0e74f4e46ce5d3c3f02c542c"} Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.125223 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6768bc7bbd-hsnsm" event={"ID":"fda7bb47-a754-4123-b031-14030f822167","Type":"ContainerDied","Data":"adcef4eee1ecba4ee0f6fee20c3ab70f63aa6a088c870b071fe2569666a4e7a1"} Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.125240 4998 scope.go:117] "RemoveContainer" containerID="3661a2f2fc4d9f66e967e4d98638f1e6d4087ebe0e74f4e46ce5d3c3f02c542c" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.125362 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6768bc7bbd-hsnsm" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.160872 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda7bb47-a754-4123-b031-14030f822167-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fda7bb47-a754-4123-b031-14030f822167" (UID: "fda7bb47-a754-4123-b031-14030f822167"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.173118 4998 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fda7bb47-a754-4123-b031-14030f822167-logs\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.173139 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q69g9\" (UniqueName: \"kubernetes.io/projected/fda7bb47-a754-4123-b031-14030f822167-kube-api-access-q69g9\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.173149 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fda7bb47-a754-4123-b031-14030f822167-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.173157 4998 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fda7bb47-a754-4123-b031-14030f822167-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.180670 4998 scope.go:117] "RemoveContainer" containerID="9d3f8d796e12715831aedb33ba672326901694d1c6e8c35f8ebcf992ac8abcdd" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.203302 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda7bb47-a754-4123-b031-14030f822167-config-data" (OuterVolumeSpecName: "config-data") pod "fda7bb47-a754-4123-b031-14030f822167" (UID: "fda7bb47-a754-4123-b031-14030f822167"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.236260 4998 scope.go:117] "RemoveContainer" containerID="3661a2f2fc4d9f66e967e4d98638f1e6d4087ebe0e74f4e46ce5d3c3f02c542c" Dec 03 16:25:30 crc kubenswrapper[4998]: E1203 16:25:30.241364 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3661a2f2fc4d9f66e967e4d98638f1e6d4087ebe0e74f4e46ce5d3c3f02c542c\": container with ID starting with 3661a2f2fc4d9f66e967e4d98638f1e6d4087ebe0e74f4e46ce5d3c3f02c542c not found: ID does not exist" containerID="3661a2f2fc4d9f66e967e4d98638f1e6d4087ebe0e74f4e46ce5d3c3f02c542c" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.241414 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3661a2f2fc4d9f66e967e4d98638f1e6d4087ebe0e74f4e46ce5d3c3f02c542c"} err="failed to get container status \"3661a2f2fc4d9f66e967e4d98638f1e6d4087ebe0e74f4e46ce5d3c3f02c542c\": rpc error: code = NotFound desc = could not find container \"3661a2f2fc4d9f66e967e4d98638f1e6d4087ebe0e74f4e46ce5d3c3f02c542c\": container with ID starting with 3661a2f2fc4d9f66e967e4d98638f1e6d4087ebe0e74f4e46ce5d3c3f02c542c not found: ID does not exist" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.241445 4998 scope.go:117] "RemoveContainer" containerID="9d3f8d796e12715831aedb33ba672326901694d1c6e8c35f8ebcf992ac8abcdd" Dec 03 16:25:30 crc kubenswrapper[4998]: E1203 16:25:30.243034 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d3f8d796e12715831aedb33ba672326901694d1c6e8c35f8ebcf992ac8abcdd\": container with ID starting with 9d3f8d796e12715831aedb33ba672326901694d1c6e8c35f8ebcf992ac8abcdd not found: ID does not exist" containerID="9d3f8d796e12715831aedb33ba672326901694d1c6e8c35f8ebcf992ac8abcdd" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.243071 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d3f8d796e12715831aedb33ba672326901694d1c6e8c35f8ebcf992ac8abcdd"} err="failed to get container status \"9d3f8d796e12715831aedb33ba672326901694d1c6e8c35f8ebcf992ac8abcdd\": rpc error: code = NotFound desc = could not find container \"9d3f8d796e12715831aedb33ba672326901694d1c6e8c35f8ebcf992ac8abcdd\": container with ID starting with 9d3f8d796e12715831aedb33ba672326901694d1c6e8c35f8ebcf992ac8abcdd not found: ID does not exist" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.277537 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fda7bb47-a754-4123-b031-14030f822167-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.336914 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55646d875-2lpr8"] Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.390891 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85c94b455f-r5qz6"] Dec 03 16:25:30 crc kubenswrapper[4998]: E1203 16:25:30.391329 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae" containerName="init" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.391341 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae" containerName="init" Dec 03 16:25:30 crc kubenswrapper[4998]: E1203 16:25:30.391359 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fda7bb47-a754-4123-b031-14030f822167" containerName="barbican-api" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.391365 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="fda7bb47-a754-4123-b031-14030f822167" containerName="barbican-api" Dec 03 16:25:30 crc kubenswrapper[4998]: E1203 16:25:30.391390 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae" containerName="dnsmasq-dns" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.391396 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae" containerName="dnsmasq-dns" Dec 03 16:25:30 crc kubenswrapper[4998]: E1203 16:25:30.391408 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de77cff6-7039-418c-8114-3c64ea4ce8e4" containerName="glance-db-sync" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.391413 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="de77cff6-7039-418c-8114-3c64ea4ce8e4" containerName="glance-db-sync" Dec 03 16:25:30 crc kubenswrapper[4998]: E1203 16:25:30.391425 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fda7bb47-a754-4123-b031-14030f822167" containerName="barbican-api-log" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.391430 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="fda7bb47-a754-4123-b031-14030f822167" containerName="barbican-api-log" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.391612 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb3db5a5-2b16-4cfc-b348-68d6a9d1cdae" containerName="dnsmasq-dns" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.391631 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="fda7bb47-a754-4123-b031-14030f822167" containerName="barbican-api-log" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.391647 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="fda7bb47-a754-4123-b031-14030f822167" containerName="barbican-api" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.391658 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="de77cff6-7039-418c-8114-3c64ea4ce8e4" containerName="glance-db-sync" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.392599 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.440231 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85c94b455f-r5qz6"] Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.483926 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-dns-svc\") pod \"dnsmasq-dns-85c94b455f-r5qz6\" (UID: \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\") " pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.483975 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfw27\" (UniqueName: \"kubernetes.io/projected/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-kube-api-access-tfw27\") pod \"dnsmasq-dns-85c94b455f-r5qz6\" (UID: \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\") " pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.484054 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-config\") pod \"dnsmasq-dns-85c94b455f-r5qz6\" (UID: \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\") " pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.484160 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-ovsdbserver-sb\") pod \"dnsmasq-dns-85c94b455f-r5qz6\" (UID: \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\") " pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.484204 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-ovsdbserver-nb\") pod \"dnsmasq-dns-85c94b455f-r5qz6\" (UID: \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\") " pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.484247 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-dns-swift-storage-0\") pod \"dnsmasq-dns-85c94b455f-r5qz6\" (UID: \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\") " pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.501860 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6768bc7bbd-hsnsm"] Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.521054 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6768bc7bbd-hsnsm"] Dec 03 16:25:30 crc kubenswrapper[4998]: E1203 16:25:30.545550 4998 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Dec 03 16:25:30 crc kubenswrapper[4998]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/72f65fee-d135-44e7-92c5-4f1afb25c2ab/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 03 16:25:30 crc kubenswrapper[4998]: > podSandboxID="e200a60c2efa9938c91ad44f5024e971c63a2c15fbb13bcb36c942aba3869614" Dec 03 16:25:30 crc kubenswrapper[4998]: E1203 16:25:30.545704 4998 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 03 16:25:30 crc kubenswrapper[4998]: container &Container{Name:dnsmasq-dns,Image:38.102.83.32:5001/podified-master-centos10/openstack-neutron-server:watcher_latest,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n598h74h59ch8ch64h599hf9hf7h668hdch8ch597h65bh59ch8dh6hc7h86h57fh649h75h586h655h57fh58bh54dh564h5b8h68fh54bh55h56dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-swift-storage-0,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-swift-storage-0,SubPath:dns-swift-storage-0,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mfqrv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-55646d875-2lpr8_openstack(72f65fee-d135-44e7-92c5-4f1afb25c2ab): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/72f65fee-d135-44e7-92c5-4f1afb25c2ab/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 03 16:25:30 crc kubenswrapper[4998]: > logger="UnhandledError" Dec 03 16:25:30 crc kubenswrapper[4998]: E1203 16:25:30.550393 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/72f65fee-d135-44e7-92c5-4f1afb25c2ab/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-55646d875-2lpr8" podUID="72f65fee-d135-44e7-92c5-4f1afb25c2ab" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.585790 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-dns-svc\") pod \"dnsmasq-dns-85c94b455f-r5qz6\" (UID: \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\") " pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.586048 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfw27\" (UniqueName: \"kubernetes.io/projected/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-kube-api-access-tfw27\") pod \"dnsmasq-dns-85c94b455f-r5qz6\" (UID: \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\") " pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.586102 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-config\") pod \"dnsmasq-dns-85c94b455f-r5qz6\" (UID: \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\") " pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.586142 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-ovsdbserver-sb\") pod \"dnsmasq-dns-85c94b455f-r5qz6\" (UID: \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\") " pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.586172 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-ovsdbserver-nb\") pod \"dnsmasq-dns-85c94b455f-r5qz6\" (UID: \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\") " pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.586204 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-dns-swift-storage-0\") pod \"dnsmasq-dns-85c94b455f-r5qz6\" (UID: \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\") " pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.587706 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-config\") pod \"dnsmasq-dns-85c94b455f-r5qz6\" (UID: \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\") " pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.588841 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-ovsdbserver-sb\") pod \"dnsmasq-dns-85c94b455f-r5qz6\" (UID: \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\") " pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.588849 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-dns-svc\") pod \"dnsmasq-dns-85c94b455f-r5qz6\" (UID: \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\") " pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.589305 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-dns-swift-storage-0\") pod \"dnsmasq-dns-85c94b455f-r5qz6\" (UID: \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\") " pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.592850 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-ovsdbserver-nb\") pod \"dnsmasq-dns-85c94b455f-r5qz6\" (UID: \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\") " pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.609472 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfw27\" (UniqueName: \"kubernetes.io/projected/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-kube-api-access-tfw27\") pod \"dnsmasq-dns-85c94b455f-r5qz6\" (UID: \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\") " pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.688128 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-77c59f659f-jkw4x"] Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.692828 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77c59f659f-jkw4x" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.703474 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.703533 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-77c59f659f-jkw4x"] Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.703656 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.720402 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.755279 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.790849 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzspm\" (UniqueName: \"kubernetes.io/projected/3aa7afc0-1cf1-4e6a-ae4e-98b13411094c-kube-api-access-xzspm\") pod \"neutron-77c59f659f-jkw4x\" (UID: \"3aa7afc0-1cf1-4e6a-ae4e-98b13411094c\") " pod="openstack/neutron-77c59f659f-jkw4x" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.791097 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3aa7afc0-1cf1-4e6a-ae4e-98b13411094c-config\") pod \"neutron-77c59f659f-jkw4x\" (UID: \"3aa7afc0-1cf1-4e6a-ae4e-98b13411094c\") " pod="openstack/neutron-77c59f659f-jkw4x" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.791187 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3aa7afc0-1cf1-4e6a-ae4e-98b13411094c-httpd-config\") pod \"neutron-77c59f659f-jkw4x\" (UID: \"3aa7afc0-1cf1-4e6a-ae4e-98b13411094c\") " pod="openstack/neutron-77c59f659f-jkw4x" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.791280 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aa7afc0-1cf1-4e6a-ae4e-98b13411094c-internal-tls-certs\") pod \"neutron-77c59f659f-jkw4x\" (UID: \"3aa7afc0-1cf1-4e6a-ae4e-98b13411094c\") " pod="openstack/neutron-77c59f659f-jkw4x" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.791356 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa7afc0-1cf1-4e6a-ae4e-98b13411094c-combined-ca-bundle\") pod \"neutron-77c59f659f-jkw4x\" (UID: \"3aa7afc0-1cf1-4e6a-ae4e-98b13411094c\") " pod="openstack/neutron-77c59f659f-jkw4x" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.791479 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aa7afc0-1cf1-4e6a-ae4e-98b13411094c-public-tls-certs\") pod \"neutron-77c59f659f-jkw4x\" (UID: \"3aa7afc0-1cf1-4e6a-ae4e-98b13411094c\") " pod="openstack/neutron-77c59f659f-jkw4x" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.791549 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aa7afc0-1cf1-4e6a-ae4e-98b13411094c-ovndb-tls-certs\") pod \"neutron-77c59f659f-jkw4x\" (UID: \"3aa7afc0-1cf1-4e6a-ae4e-98b13411094c\") " pod="openstack/neutron-77c59f659f-jkw4x" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.804249 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.895891 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzspm\" (UniqueName: \"kubernetes.io/projected/3aa7afc0-1cf1-4e6a-ae4e-98b13411094c-kube-api-access-xzspm\") pod \"neutron-77c59f659f-jkw4x\" (UID: \"3aa7afc0-1cf1-4e6a-ae4e-98b13411094c\") " pod="openstack/neutron-77c59f659f-jkw4x" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.895959 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3aa7afc0-1cf1-4e6a-ae4e-98b13411094c-config\") pod \"neutron-77c59f659f-jkw4x\" (UID: \"3aa7afc0-1cf1-4e6a-ae4e-98b13411094c\") " pod="openstack/neutron-77c59f659f-jkw4x" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.895983 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3aa7afc0-1cf1-4e6a-ae4e-98b13411094c-httpd-config\") pod \"neutron-77c59f659f-jkw4x\" (UID: \"3aa7afc0-1cf1-4e6a-ae4e-98b13411094c\") " pod="openstack/neutron-77c59f659f-jkw4x" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.896018 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aa7afc0-1cf1-4e6a-ae4e-98b13411094c-internal-tls-certs\") pod \"neutron-77c59f659f-jkw4x\" (UID: \"3aa7afc0-1cf1-4e6a-ae4e-98b13411094c\") " pod="openstack/neutron-77c59f659f-jkw4x" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.896039 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa7afc0-1cf1-4e6a-ae4e-98b13411094c-combined-ca-bundle\") pod \"neutron-77c59f659f-jkw4x\" (UID: \"3aa7afc0-1cf1-4e6a-ae4e-98b13411094c\") " pod="openstack/neutron-77c59f659f-jkw4x" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.896124 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aa7afc0-1cf1-4e6a-ae4e-98b13411094c-public-tls-certs\") pod \"neutron-77c59f659f-jkw4x\" (UID: \"3aa7afc0-1cf1-4e6a-ae4e-98b13411094c\") " pod="openstack/neutron-77c59f659f-jkw4x" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.896142 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aa7afc0-1cf1-4e6a-ae4e-98b13411094c-ovndb-tls-certs\") pod \"neutron-77c59f659f-jkw4x\" (UID: \"3aa7afc0-1cf1-4e6a-ae4e-98b13411094c\") " pod="openstack/neutron-77c59f659f-jkw4x" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.908447 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aa7afc0-1cf1-4e6a-ae4e-98b13411094c-internal-tls-certs\") pod \"neutron-77c59f659f-jkw4x\" (UID: \"3aa7afc0-1cf1-4e6a-ae4e-98b13411094c\") " pod="openstack/neutron-77c59f659f-jkw4x" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.910677 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aa7afc0-1cf1-4e6a-ae4e-98b13411094c-public-tls-certs\") pod \"neutron-77c59f659f-jkw4x\" (UID: \"3aa7afc0-1cf1-4e6a-ae4e-98b13411094c\") " pod="openstack/neutron-77c59f659f-jkw4x" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.911466 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3aa7afc0-1cf1-4e6a-ae4e-98b13411094c-ovndb-tls-certs\") pod \"neutron-77c59f659f-jkw4x\" (UID: \"3aa7afc0-1cf1-4e6a-ae4e-98b13411094c\") " pod="openstack/neutron-77c59f659f-jkw4x" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.916767 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3aa7afc0-1cf1-4e6a-ae4e-98b13411094c-config\") pod \"neutron-77c59f659f-jkw4x\" (UID: \"3aa7afc0-1cf1-4e6a-ae4e-98b13411094c\") " pod="openstack/neutron-77c59f659f-jkw4x" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.920674 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa7afc0-1cf1-4e6a-ae4e-98b13411094c-combined-ca-bundle\") pod \"neutron-77c59f659f-jkw4x\" (UID: \"3aa7afc0-1cf1-4e6a-ae4e-98b13411094c\") " pod="openstack/neutron-77c59f659f-jkw4x" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.921312 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3aa7afc0-1cf1-4e6a-ae4e-98b13411094c-httpd-config\") pod \"neutron-77c59f659f-jkw4x\" (UID: \"3aa7afc0-1cf1-4e6a-ae4e-98b13411094c\") " pod="openstack/neutron-77c59f659f-jkw4x" Dec 03 16:25:30 crc kubenswrapper[4998]: I1203 16:25:30.925554 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzspm\" (UniqueName: \"kubernetes.io/projected/3aa7afc0-1cf1-4e6a-ae4e-98b13411094c-kube-api-access-xzspm\") pod \"neutron-77c59f659f-jkw4x\" (UID: \"3aa7afc0-1cf1-4e6a-ae4e-98b13411094c\") " pod="openstack/neutron-77c59f659f-jkw4x" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.038193 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77c59f659f-jkw4x" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.244379 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f47995564-5mql2" event={"ID":"98bb0f8f-aa52-4408-8d16-c6677b080101","Type":"ContainerStarted","Data":"a1d30b67a1dfea698ecb9b1bea63314a38f9de7898fc55974c57c35c76092c39"} Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.245035 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7f47995564-5mql2" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.258842 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e6872e78-e3b5-4024-9de0-24c6a3cfe345" containerName="cinder-scheduler" containerID="cri-o://47ad5ff218d1db4e4a839220fd027f05fd704ae47e711f9827767480718bfd25" gracePeriod=30 Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.259285 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e6872e78-e3b5-4024-9de0-24c6a3cfe345" containerName="probe" containerID="cri-o://9a133c288c85b51f5f5a60e0fe7913b20f9fd4a9bf8104561e23e33479dada46" gracePeriod=30 Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.365296 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7f47995564-5mql2" podStartSLOduration=3.36527352 podStartE2EDuration="3.36527352s" podCreationTimestamp="2025-12-03 16:25:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:25:31.283681476 +0000 UTC m=+1309.895381719" watchObservedRunningTime="2025-12-03 16:25:31.36527352 +0000 UTC m=+1309.976973743" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.430816 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.473723 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.474494 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.479092 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bpb5n" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.479323 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.487752 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85c94b455f-r5qz6"] Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.489395 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.596807 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.598800 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.602241 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.609079 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.673281 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ee49423-2774-4d51-ae4f-8cea1775efef-scripts\") pod \"glance-default-external-api-0\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.673415 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqf6x\" (UniqueName: \"kubernetes.io/projected/1ee49423-2774-4d51-ae4f-8cea1775efef-kube-api-access-sqf6x\") pod \"glance-default-external-api-0\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.673532 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ee49423-2774-4d51-ae4f-8cea1775efef-config-data\") pod \"glance-default-external-api-0\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.673580 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ee49423-2774-4d51-ae4f-8cea1775efef-logs\") pod \"glance-default-external-api-0\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.673604 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ee49423-2774-4d51-ae4f-8cea1775efef-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.673657 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.673693 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1ee49423-2774-4d51-ae4f-8cea1775efef-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.732187 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda7bb47-a754-4123-b031-14030f822167" path="/var/lib/kubelet/pods/fda7bb47-a754-4123-b031-14030f822167/volumes" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.775801 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1ee49423-2774-4d51-ae4f-8cea1775efef-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.775861 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f32474ed-1700-42c9-b87d-ffc1f8445ea6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.775878 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f32474ed-1700-42c9-b87d-ffc1f8445ea6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.775893 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnpp7\" (UniqueName: \"kubernetes.io/projected/f32474ed-1700-42c9-b87d-ffc1f8445ea6-kube-api-access-lnpp7\") pod \"glance-default-internal-api-0\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.775928 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f32474ed-1700-42c9-b87d-ffc1f8445ea6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.775954 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ee49423-2774-4d51-ae4f-8cea1775efef-scripts\") pod \"glance-default-external-api-0\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.776007 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqf6x\" (UniqueName: \"kubernetes.io/projected/1ee49423-2774-4d51-ae4f-8cea1775efef-kube-api-access-sqf6x\") pod \"glance-default-external-api-0\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.776057 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f32474ed-1700-42c9-b87d-ffc1f8445ea6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.776086 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.776108 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ee49423-2774-4d51-ae4f-8cea1775efef-config-data\") pod \"glance-default-external-api-0\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.776138 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ee49423-2774-4d51-ae4f-8cea1775efef-logs\") pod \"glance-default-external-api-0\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.776156 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ee49423-2774-4d51-ae4f-8cea1775efef-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.776182 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f32474ed-1700-42c9-b87d-ffc1f8445ea6-logs\") pod \"glance-default-internal-api-0\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.776199 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.776606 4998 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.779741 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1ee49423-2774-4d51-ae4f-8cea1775efef-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.782777 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ee49423-2774-4d51-ae4f-8cea1775efef-logs\") pod \"glance-default-external-api-0\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.782984 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55646d875-2lpr8" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.799881 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ee49423-2774-4d51-ae4f-8cea1775efef-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.809095 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqf6x\" (UniqueName: \"kubernetes.io/projected/1ee49423-2774-4d51-ae4f-8cea1775efef-kube-api-access-sqf6x\") pod \"glance-default-external-api-0\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.809794 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ee49423-2774-4d51-ae4f-8cea1775efef-scripts\") pod \"glance-default-external-api-0\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.811777 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ee49423-2774-4d51-ae4f-8cea1775efef-config-data\") pod \"glance-default-external-api-0\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.874008 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.877601 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-dns-svc\") pod \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\" (UID: \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\") " Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.877776 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfqrv\" (UniqueName: \"kubernetes.io/projected/72f65fee-d135-44e7-92c5-4f1afb25c2ab-kube-api-access-mfqrv\") pod \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\" (UID: \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\") " Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.877926 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-dns-swift-storage-0\") pod \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\" (UID: \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\") " Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.877968 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-config\") pod \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\" (UID: \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\") " Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.878029 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-ovsdbserver-nb\") pod \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\" (UID: \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\") " Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.878104 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-ovsdbserver-sb\") pod \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\" (UID: \"72f65fee-d135-44e7-92c5-4f1afb25c2ab\") " Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.878362 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f32474ed-1700-42c9-b87d-ffc1f8445ea6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.878394 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.878432 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f32474ed-1700-42c9-b87d-ffc1f8445ea6-logs\") pod \"glance-default-internal-api-0\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.878472 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f32474ed-1700-42c9-b87d-ffc1f8445ea6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.878487 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f32474ed-1700-42c9-b87d-ffc1f8445ea6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.878502 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnpp7\" (UniqueName: \"kubernetes.io/projected/f32474ed-1700-42c9-b87d-ffc1f8445ea6-kube-api-access-lnpp7\") pod \"glance-default-internal-api-0\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.878769 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f32474ed-1700-42c9-b87d-ffc1f8445ea6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.880208 4998 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.882233 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f32474ed-1700-42c9-b87d-ffc1f8445ea6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.887090 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f32474ed-1700-42c9-b87d-ffc1f8445ea6-logs\") pod \"glance-default-internal-api-0\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.888411 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f32474ed-1700-42c9-b87d-ffc1f8445ea6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.896259 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f32474ed-1700-42c9-b87d-ffc1f8445ea6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.905724 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72f65fee-d135-44e7-92c5-4f1afb25c2ab-kube-api-access-mfqrv" (OuterVolumeSpecName: "kube-api-access-mfqrv") pod "72f65fee-d135-44e7-92c5-4f1afb25c2ab" (UID: "72f65fee-d135-44e7-92c5-4f1afb25c2ab"). InnerVolumeSpecName "kube-api-access-mfqrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.905810 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f32474ed-1700-42c9-b87d-ffc1f8445ea6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.949559 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnpp7\" (UniqueName: \"kubernetes.io/projected/f32474ed-1700-42c9-b87d-ffc1f8445ea6-kube-api-access-lnpp7\") pod \"glance-default-internal-api-0\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.965325 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.981015 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfqrv\" (UniqueName: \"kubernetes.io/projected/72f65fee-d135-44e7-92c5-4f1afb25c2ab-kube-api-access-mfqrv\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:31 crc kubenswrapper[4998]: I1203 16:25:31.989017 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-77c59f659f-jkw4x"] Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.012524 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.048676 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.075734 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "72f65fee-d135-44e7-92c5-4f1afb25c2ab" (UID: "72f65fee-d135-44e7-92c5-4f1afb25c2ab"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.079417 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "72f65fee-d135-44e7-92c5-4f1afb25c2ab" (UID: "72f65fee-d135-44e7-92c5-4f1afb25c2ab"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.088831 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lt2b\" (UniqueName: \"kubernetes.io/projected/d7600639-7175-4c0a-9524-40adb505d3ea-kube-api-access-2lt2b\") pod \"d7600639-7175-4c0a-9524-40adb505d3ea\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.090441 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7600639-7175-4c0a-9524-40adb505d3ea-logs\") pod \"d7600639-7175-4c0a-9524-40adb505d3ea\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.090568 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7600639-7175-4c0a-9524-40adb505d3ea-combined-ca-bundle\") pod \"d7600639-7175-4c0a-9524-40adb505d3ea\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.090670 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7600639-7175-4c0a-9524-40adb505d3ea-scripts\") pod \"d7600639-7175-4c0a-9524-40adb505d3ea\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.090712 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d7600639-7175-4c0a-9524-40adb505d3ea-horizon-secret-key\") pod \"d7600639-7175-4c0a-9524-40adb505d3ea\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.091062 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7600639-7175-4c0a-9524-40adb505d3ea-horizon-tls-certs\") pod \"d7600639-7175-4c0a-9524-40adb505d3ea\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.091120 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d7600639-7175-4c0a-9524-40adb505d3ea-config-data\") pod \"d7600639-7175-4c0a-9524-40adb505d3ea\" (UID: \"d7600639-7175-4c0a-9524-40adb505d3ea\") " Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.091617 4998 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.091632 4998 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.105606 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7600639-7175-4c0a-9524-40adb505d3ea-logs" (OuterVolumeSpecName: "logs") pod "d7600639-7175-4c0a-9524-40adb505d3ea" (UID: "d7600639-7175-4c0a-9524-40adb505d3ea"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.105627 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7600639-7175-4c0a-9524-40adb505d3ea-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d7600639-7175-4c0a-9524-40adb505d3ea" (UID: "d7600639-7175-4c0a-9524-40adb505d3ea"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.109028 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7600639-7175-4c0a-9524-40adb505d3ea-kube-api-access-2lt2b" (OuterVolumeSpecName: "kube-api-access-2lt2b") pod "d7600639-7175-4c0a-9524-40adb505d3ea" (UID: "d7600639-7175-4c0a-9524-40adb505d3ea"). InnerVolumeSpecName "kube-api-access-2lt2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.114002 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-config" (OuterVolumeSpecName: "config") pod "72f65fee-d135-44e7-92c5-4f1afb25c2ab" (UID: "72f65fee-d135-44e7-92c5-4f1afb25c2ab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.120671 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7600639-7175-4c0a-9524-40adb505d3ea-config-data" (OuterVolumeSpecName: "config-data") pod "d7600639-7175-4c0a-9524-40adb505d3ea" (UID: "d7600639-7175-4c0a-9524-40adb505d3ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.136616 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "72f65fee-d135-44e7-92c5-4f1afb25c2ab" (UID: "72f65fee-d135-44e7-92c5-4f1afb25c2ab"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.142954 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7600639-7175-4c0a-9524-40adb505d3ea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7600639-7175-4c0a-9524-40adb505d3ea" (UID: "d7600639-7175-4c0a-9524-40adb505d3ea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.151315 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7600639-7175-4c0a-9524-40adb505d3ea-scripts" (OuterVolumeSpecName: "scripts") pod "d7600639-7175-4c0a-9524-40adb505d3ea" (UID: "d7600639-7175-4c0a-9524-40adb505d3ea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.155337 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "72f65fee-d135-44e7-92c5-4f1afb25c2ab" (UID: "72f65fee-d135-44e7-92c5-4f1afb25c2ab"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.170731 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7600639-7175-4c0a-9524-40adb505d3ea-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "d7600639-7175-4c0a-9524-40adb505d3ea" (UID: "d7600639-7175-4c0a-9524-40adb505d3ea"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.193807 4998 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.193834 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7600639-7175-4c0a-9524-40adb505d3ea-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.193843 4998 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d7600639-7175-4c0a-9524-40adb505d3ea-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.193854 4998 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.193862 4998 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7600639-7175-4c0a-9524-40adb505d3ea-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.193870 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d7600639-7175-4c0a-9524-40adb505d3ea-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.193878 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lt2b\" (UniqueName: \"kubernetes.io/projected/d7600639-7175-4c0a-9524-40adb505d3ea-kube-api-access-2lt2b\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.193887 4998 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7600639-7175-4c0a-9524-40adb505d3ea-logs\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.193895 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72f65fee-d135-44e7-92c5-4f1afb25c2ab-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.193902 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7600639-7175-4c0a-9524-40adb505d3ea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.323312 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.340298 4998 generic.go:334] "Generic (PLEG): container finished" podID="d7600639-7175-4c0a-9524-40adb505d3ea" containerID="75a61917c7bd2bcb544cb164de1c7e5b8e8e2cbe5262c7fa8075c0db387aa835" exitCode=137 Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.340371 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7565d78f8d-jm9qt" event={"ID":"d7600639-7175-4c0a-9524-40adb505d3ea","Type":"ContainerDied","Data":"75a61917c7bd2bcb544cb164de1c7e5b8e8e2cbe5262c7fa8075c0db387aa835"} Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.340404 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7565d78f8d-jm9qt" event={"ID":"d7600639-7175-4c0a-9524-40adb505d3ea","Type":"ContainerDied","Data":"6eb3a39450e43793101cc3b0a6e6381ea10674fde4b0cc6eedd34b0de43e1c1b"} Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.340426 4998 scope.go:117] "RemoveContainer" containerID="549d68f35b182d6fa1f0327a26451914c3855dd97314b4621e055785c688e514" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.340547 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7565d78f8d-jm9qt" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.360392 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55646d875-2lpr8" event={"ID":"72f65fee-d135-44e7-92c5-4f1afb25c2ab","Type":"ContainerDied","Data":"e200a60c2efa9938c91ad44f5024e971c63a2c15fbb13bcb36c942aba3869614"} Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.360474 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55646d875-2lpr8" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.366659 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77c59f659f-jkw4x" event={"ID":"3aa7afc0-1cf1-4e6a-ae4e-98b13411094c","Type":"ContainerStarted","Data":"c597b17bb7d2d79ba6f4596c290c5b8fab8e7a305db69b10082cc0a7d8e219e3"} Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.368667 4998 generic.go:334] "Generic (PLEG): container finished" podID="e22aa31f-72c8-45c5-bcbc-91c5ada645a3" containerID="7d7328b400c9283c54c6fc3ae67e3d0c99c825f5e02471ab881dabacee8513d5" exitCode=0 Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.368943 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" event={"ID":"e22aa31f-72c8-45c5-bcbc-91c5ada645a3","Type":"ContainerDied","Data":"7d7328b400c9283c54c6fc3ae67e3d0c99c825f5e02471ab881dabacee8513d5"} Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.369075 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" event={"ID":"e22aa31f-72c8-45c5-bcbc-91c5ada645a3","Type":"ContainerStarted","Data":"4e293e6dfdf06990c533afb15796569a0e9fe89f5ee647a6bc6041e92a23f7ec"} Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.712823 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7565d78f8d-jm9qt"] Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.725909 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7565d78f8d-jm9qt"] Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.736364 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.837007 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55646d875-2lpr8"] Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.859411 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-57c877d99d-7qbcv" Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.899370 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55646d875-2lpr8"] Dec 03 16:25:32 crc kubenswrapper[4998]: I1203 16:25:32.932715 4998 scope.go:117] "RemoveContainer" containerID="75a61917c7bd2bcb544cb164de1c7e5b8e8e2cbe5262c7fa8075c0db387aa835" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.023964 4998 scope.go:117] "RemoveContainer" containerID="549d68f35b182d6fa1f0327a26451914c3855dd97314b4621e055785c688e514" Dec 03 16:25:33 crc kubenswrapper[4998]: E1203 16:25:33.028843 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"549d68f35b182d6fa1f0327a26451914c3855dd97314b4621e055785c688e514\": container with ID starting with 549d68f35b182d6fa1f0327a26451914c3855dd97314b4621e055785c688e514 not found: ID does not exist" containerID="549d68f35b182d6fa1f0327a26451914c3855dd97314b4621e055785c688e514" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.028878 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"549d68f35b182d6fa1f0327a26451914c3855dd97314b4621e055785c688e514"} err="failed to get container status \"549d68f35b182d6fa1f0327a26451914c3855dd97314b4621e055785c688e514\": rpc error: code = NotFound desc = could not find container \"549d68f35b182d6fa1f0327a26451914c3855dd97314b4621e055785c688e514\": container with ID starting with 549d68f35b182d6fa1f0327a26451914c3855dd97314b4621e055785c688e514 not found: ID does not exist" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.028898 4998 scope.go:117] "RemoveContainer" containerID="75a61917c7bd2bcb544cb164de1c7e5b8e8e2cbe5262c7fa8075c0db387aa835" Dec 03 16:25:33 crc kubenswrapper[4998]: E1203 16:25:33.033844 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75a61917c7bd2bcb544cb164de1c7e5b8e8e2cbe5262c7fa8075c0db387aa835\": container with ID starting with 75a61917c7bd2bcb544cb164de1c7e5b8e8e2cbe5262c7fa8075c0db387aa835 not found: ID does not exist" containerID="75a61917c7bd2bcb544cb164de1c7e5b8e8e2cbe5262c7fa8075c0db387aa835" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.033869 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75a61917c7bd2bcb544cb164de1c7e5b8e8e2cbe5262c7fa8075c0db387aa835"} err="failed to get container status \"75a61917c7bd2bcb544cb164de1c7e5b8e8e2cbe5262c7fa8075c0db387aa835\": rpc error: code = NotFound desc = could not find container \"75a61917c7bd2bcb544cb164de1c7e5b8e8e2cbe5262c7fa8075c0db387aa835\": container with ID starting with 75a61917c7bd2bcb544cb164de1c7e5b8e8e2cbe5262c7fa8075c0db387aa835 not found: ID does not exist" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.033884 4998 scope.go:117] "RemoveContainer" containerID="b4142ae26f57f5d78cff31bfad2f6cadffe6d46285dfcc3318fe104337bd5dc5" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.193212 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.241850 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6872e78-e3b5-4024-9de0-24c6a3cfe345-scripts\") pod \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\" (UID: \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\") " Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.242266 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6872e78-e3b5-4024-9de0-24c6a3cfe345-config-data\") pod \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\" (UID: \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\") " Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.242314 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6872e78-e3b5-4024-9de0-24c6a3cfe345-combined-ca-bundle\") pod \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\" (UID: \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\") " Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.242335 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e6872e78-e3b5-4024-9de0-24c6a3cfe345-config-data-custom\") pod \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\" (UID: \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\") " Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.242369 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66ggp\" (UniqueName: \"kubernetes.io/projected/e6872e78-e3b5-4024-9de0-24c6a3cfe345-kube-api-access-66ggp\") pod \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\" (UID: \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\") " Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.242420 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e6872e78-e3b5-4024-9de0-24c6a3cfe345-etc-machine-id\") pod \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\" (UID: \"e6872e78-e3b5-4024-9de0-24c6a3cfe345\") " Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.242885 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e6872e78-e3b5-4024-9de0-24c6a3cfe345-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e6872e78-e3b5-4024-9de0-24c6a3cfe345" (UID: "e6872e78-e3b5-4024-9de0-24c6a3cfe345"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.248171 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6872e78-e3b5-4024-9de0-24c6a3cfe345-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e6872e78-e3b5-4024-9de0-24c6a3cfe345" (UID: "e6872e78-e3b5-4024-9de0-24c6a3cfe345"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.248559 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6872e78-e3b5-4024-9de0-24c6a3cfe345-kube-api-access-66ggp" (OuterVolumeSpecName: "kube-api-access-66ggp") pod "e6872e78-e3b5-4024-9de0-24c6a3cfe345" (UID: "e6872e78-e3b5-4024-9de0-24c6a3cfe345"). InnerVolumeSpecName "kube-api-access-66ggp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.262281 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6872e78-e3b5-4024-9de0-24c6a3cfe345-scripts" (OuterVolumeSpecName: "scripts") pod "e6872e78-e3b5-4024-9de0-24c6a3cfe345" (UID: "e6872e78-e3b5-4024-9de0-24c6a3cfe345"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.354063 4998 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e6872e78-e3b5-4024-9de0-24c6a3cfe345-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.354090 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6872e78-e3b5-4024-9de0-24c6a3cfe345-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.354099 4998 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e6872e78-e3b5-4024-9de0-24c6a3cfe345-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.354108 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66ggp\" (UniqueName: \"kubernetes.io/projected/e6872e78-e3b5-4024-9de0-24c6a3cfe345-kube-api-access-66ggp\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.427052 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6872e78-e3b5-4024-9de0-24c6a3cfe345-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e6872e78-e3b5-4024-9de0-24c6a3cfe345" (UID: "e6872e78-e3b5-4024-9de0-24c6a3cfe345"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.428209 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" event={"ID":"e22aa31f-72c8-45c5-bcbc-91c5ada645a3","Type":"ContainerStarted","Data":"2af0ef43191ec1a4c35ffb9f96b92eb4a1259eaa63c5a31a2d279444fb0c33c6"} Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.428312 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.431666 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 16:25:33 crc kubenswrapper[4998]: W1203 16:25:33.434232 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf32474ed_1700_42c9_b87d_ffc1f8445ea6.slice/crio-97dd6453d32f7151528140a97eace904d6bd6eb130c780341f43e4e246d4e9a3 WatchSource:0}: Error finding container 97dd6453d32f7151528140a97eace904d6bd6eb130c780341f43e4e246d4e9a3: Status 404 returned error can't find the container with id 97dd6453d32f7151528140a97eace904d6bd6eb130c780341f43e4e246d4e9a3 Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.435989 4998 generic.go:334] "Generic (PLEG): container finished" podID="e6872e78-e3b5-4024-9de0-24c6a3cfe345" containerID="9a133c288c85b51f5f5a60e0fe7913b20f9fd4a9bf8104561e23e33479dada46" exitCode=0 Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.436010 4998 generic.go:334] "Generic (PLEG): container finished" podID="e6872e78-e3b5-4024-9de0-24c6a3cfe345" containerID="47ad5ff218d1db4e4a839220fd027f05fd704ae47e711f9827767480718bfd25" exitCode=0 Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.436043 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e6872e78-e3b5-4024-9de0-24c6a3cfe345","Type":"ContainerDied","Data":"9a133c288c85b51f5f5a60e0fe7913b20f9fd4a9bf8104561e23e33479dada46"} Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.436063 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e6872e78-e3b5-4024-9de0-24c6a3cfe345","Type":"ContainerDied","Data":"47ad5ff218d1db4e4a839220fd027f05fd704ae47e711f9827767480718bfd25"} Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.436072 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e6872e78-e3b5-4024-9de0-24c6a3cfe345","Type":"ContainerDied","Data":"6825d6228b457e2e86c5bf832da1451c4e304e994cb5de0693806ec8f3ad36d4"} Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.436086 4998 scope.go:117] "RemoveContainer" containerID="9a133c288c85b51f5f5a60e0fe7913b20f9fd4a9bf8104561e23e33479dada46" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.436168 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.440800 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1ee49423-2774-4d51-ae4f-8cea1775efef","Type":"ContainerStarted","Data":"97bbe409920beffc63a9ade541117bff7a7e66ee697bbf668c455419561bfdc4"} Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.443180 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6872e78-e3b5-4024-9de0-24c6a3cfe345-config-data" (OuterVolumeSpecName: "config-data") pod "e6872e78-e3b5-4024-9de0-24c6a3cfe345" (UID: "e6872e78-e3b5-4024-9de0-24c6a3cfe345"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.448232 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77c59f659f-jkw4x" event={"ID":"3aa7afc0-1cf1-4e6a-ae4e-98b13411094c","Type":"ContainerStarted","Data":"c67f2d84101c62721222befb9cce4393b8aed85b9622498e21c2d741192159ea"} Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.456305 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6872e78-e3b5-4024-9de0-24c6a3cfe345-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.456348 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6872e78-e3b5-4024-9de0-24c6a3cfe345-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.457414 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" podStartSLOduration=3.457381491 podStartE2EDuration="3.457381491s" podCreationTimestamp="2025-12-03 16:25:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:25:33.453338893 +0000 UTC m=+1312.065039116" watchObservedRunningTime="2025-12-03 16:25:33.457381491 +0000 UTC m=+1312.069081714" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.584926 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-59d5c55c95-qdh4t"] Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.597915 4998 scope.go:117] "RemoveContainer" containerID="47ad5ff218d1db4e4a839220fd027f05fd704ae47e711f9827767480718bfd25" Dec 03 16:25:33 crc kubenswrapper[4998]: E1203 16:25:33.612936 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6872e78-e3b5-4024-9de0-24c6a3cfe345" containerName="probe" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.612973 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6872e78-e3b5-4024-9de0-24c6a3cfe345" containerName="probe" Dec 03 16:25:33 crc kubenswrapper[4998]: E1203 16:25:33.612987 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7600639-7175-4c0a-9524-40adb505d3ea" containerName="horizon-log" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.612992 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7600639-7175-4c0a-9524-40adb505d3ea" containerName="horizon-log" Dec 03 16:25:33 crc kubenswrapper[4998]: E1203 16:25:33.613012 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6872e78-e3b5-4024-9de0-24c6a3cfe345" containerName="cinder-scheduler" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.613022 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6872e78-e3b5-4024-9de0-24c6a3cfe345" containerName="cinder-scheduler" Dec 03 16:25:33 crc kubenswrapper[4998]: E1203 16:25:33.613046 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7600639-7175-4c0a-9524-40adb505d3ea" containerName="horizon" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.613052 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7600639-7175-4c0a-9524-40adb505d3ea" containerName="horizon" Dec 03 16:25:33 crc kubenswrapper[4998]: E1203 16:25:33.613066 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72f65fee-d135-44e7-92c5-4f1afb25c2ab" containerName="init" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.613071 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="72f65fee-d135-44e7-92c5-4f1afb25c2ab" containerName="init" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.613328 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6872e78-e3b5-4024-9de0-24c6a3cfe345" containerName="probe" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.613338 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="72f65fee-d135-44e7-92c5-4f1afb25c2ab" containerName="init" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.613347 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7600639-7175-4c0a-9524-40adb505d3ea" containerName="horizon-log" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.613356 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7600639-7175-4c0a-9524-40adb505d3ea" containerName="horizon" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.613367 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6872e78-e3b5-4024-9de0-24c6a3cfe345" containerName="cinder-scheduler" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.614358 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.638832 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.639032 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.639193 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.665293 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-59d5c55c95-qdh4t"] Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.674723 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cc160003-2b7c-4772-a0f3-7b431c501fb8-etc-swift\") pod \"swift-proxy-59d5c55c95-qdh4t\" (UID: \"cc160003-2b7c-4772-a0f3-7b431c501fb8\") " pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.704445 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc160003-2b7c-4772-a0f3-7b431c501fb8-config-data\") pod \"swift-proxy-59d5c55c95-qdh4t\" (UID: \"cc160003-2b7c-4772-a0f3-7b431c501fb8\") " pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.704504 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cc160003-2b7c-4772-a0f3-7b431c501fb8-log-httpd\") pod \"swift-proxy-59d5c55c95-qdh4t\" (UID: \"cc160003-2b7c-4772-a0f3-7b431c501fb8\") " pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.704522 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xm7l\" (UniqueName: \"kubernetes.io/projected/cc160003-2b7c-4772-a0f3-7b431c501fb8-kube-api-access-9xm7l\") pod \"swift-proxy-59d5c55c95-qdh4t\" (UID: \"cc160003-2b7c-4772-a0f3-7b431c501fb8\") " pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.704621 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cc160003-2b7c-4772-a0f3-7b431c501fb8-run-httpd\") pod \"swift-proxy-59d5c55c95-qdh4t\" (UID: \"cc160003-2b7c-4772-a0f3-7b431c501fb8\") " pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.704709 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc160003-2b7c-4772-a0f3-7b431c501fb8-internal-tls-certs\") pod \"swift-proxy-59d5c55c95-qdh4t\" (UID: \"cc160003-2b7c-4772-a0f3-7b431c501fb8\") " pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.704867 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc160003-2b7c-4772-a0f3-7b431c501fb8-public-tls-certs\") pod \"swift-proxy-59d5c55c95-qdh4t\" (UID: \"cc160003-2b7c-4772-a0f3-7b431c501fb8\") " pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.704938 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc160003-2b7c-4772-a0f3-7b431c501fb8-combined-ca-bundle\") pod \"swift-proxy-59d5c55c95-qdh4t\" (UID: \"cc160003-2b7c-4772-a0f3-7b431c501fb8\") " pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.739057 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72f65fee-d135-44e7-92c5-4f1afb25c2ab" path="/var/lib/kubelet/pods/72f65fee-d135-44e7-92c5-4f1afb25c2ab/volumes" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.739697 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7600639-7175-4c0a-9524-40adb505d3ea" path="/var/lib/kubelet/pods/d7600639-7175-4c0a-9524-40adb505d3ea/volumes" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.815950 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc160003-2b7c-4772-a0f3-7b431c501fb8-internal-tls-certs\") pod \"swift-proxy-59d5c55c95-qdh4t\" (UID: \"cc160003-2b7c-4772-a0f3-7b431c501fb8\") " pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.816035 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc160003-2b7c-4772-a0f3-7b431c501fb8-public-tls-certs\") pod \"swift-proxy-59d5c55c95-qdh4t\" (UID: \"cc160003-2b7c-4772-a0f3-7b431c501fb8\") " pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.816067 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc160003-2b7c-4772-a0f3-7b431c501fb8-combined-ca-bundle\") pod \"swift-proxy-59d5c55c95-qdh4t\" (UID: \"cc160003-2b7c-4772-a0f3-7b431c501fb8\") " pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.816111 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cc160003-2b7c-4772-a0f3-7b431c501fb8-etc-swift\") pod \"swift-proxy-59d5c55c95-qdh4t\" (UID: \"cc160003-2b7c-4772-a0f3-7b431c501fb8\") " pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.816155 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc160003-2b7c-4772-a0f3-7b431c501fb8-config-data\") pod \"swift-proxy-59d5c55c95-qdh4t\" (UID: \"cc160003-2b7c-4772-a0f3-7b431c501fb8\") " pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.816179 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cc160003-2b7c-4772-a0f3-7b431c501fb8-log-httpd\") pod \"swift-proxy-59d5c55c95-qdh4t\" (UID: \"cc160003-2b7c-4772-a0f3-7b431c501fb8\") " pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.816199 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xm7l\" (UniqueName: \"kubernetes.io/projected/cc160003-2b7c-4772-a0f3-7b431c501fb8-kube-api-access-9xm7l\") pod \"swift-proxy-59d5c55c95-qdh4t\" (UID: \"cc160003-2b7c-4772-a0f3-7b431c501fb8\") " pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.816255 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cc160003-2b7c-4772-a0f3-7b431c501fb8-run-httpd\") pod \"swift-proxy-59d5c55c95-qdh4t\" (UID: \"cc160003-2b7c-4772-a0f3-7b431c501fb8\") " pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.816705 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cc160003-2b7c-4772-a0f3-7b431c501fb8-run-httpd\") pod \"swift-proxy-59d5c55c95-qdh4t\" (UID: \"cc160003-2b7c-4772-a0f3-7b431c501fb8\") " pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.817719 4998 scope.go:117] "RemoveContainer" containerID="9a133c288c85b51f5f5a60e0fe7913b20f9fd4a9bf8104561e23e33479dada46" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.819003 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cc160003-2b7c-4772-a0f3-7b431c501fb8-log-httpd\") pod \"swift-proxy-59d5c55c95-qdh4t\" (UID: \"cc160003-2b7c-4772-a0f3-7b431c501fb8\") " pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:33 crc kubenswrapper[4998]: E1203 16:25:33.833625 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a133c288c85b51f5f5a60e0fe7913b20f9fd4a9bf8104561e23e33479dada46\": container with ID starting with 9a133c288c85b51f5f5a60e0fe7913b20f9fd4a9bf8104561e23e33479dada46 not found: ID does not exist" containerID="9a133c288c85b51f5f5a60e0fe7913b20f9fd4a9bf8104561e23e33479dada46" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.833675 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a133c288c85b51f5f5a60e0fe7913b20f9fd4a9bf8104561e23e33479dada46"} err="failed to get container status \"9a133c288c85b51f5f5a60e0fe7913b20f9fd4a9bf8104561e23e33479dada46\": rpc error: code = NotFound desc = could not find container \"9a133c288c85b51f5f5a60e0fe7913b20f9fd4a9bf8104561e23e33479dada46\": container with ID starting with 9a133c288c85b51f5f5a60e0fe7913b20f9fd4a9bf8104561e23e33479dada46 not found: ID does not exist" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.833702 4998 scope.go:117] "RemoveContainer" containerID="47ad5ff218d1db4e4a839220fd027f05fd704ae47e711f9827767480718bfd25" Dec 03 16:25:33 crc kubenswrapper[4998]: E1203 16:25:33.834949 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47ad5ff218d1db4e4a839220fd027f05fd704ae47e711f9827767480718bfd25\": container with ID starting with 47ad5ff218d1db4e4a839220fd027f05fd704ae47e711f9827767480718bfd25 not found: ID does not exist" containerID="47ad5ff218d1db4e4a839220fd027f05fd704ae47e711f9827767480718bfd25" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.835001 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47ad5ff218d1db4e4a839220fd027f05fd704ae47e711f9827767480718bfd25"} err="failed to get container status \"47ad5ff218d1db4e4a839220fd027f05fd704ae47e711f9827767480718bfd25\": rpc error: code = NotFound desc = could not find container \"47ad5ff218d1db4e4a839220fd027f05fd704ae47e711f9827767480718bfd25\": container with ID starting with 47ad5ff218d1db4e4a839220fd027f05fd704ae47e711f9827767480718bfd25 not found: ID does not exist" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.835035 4998 scope.go:117] "RemoveContainer" containerID="9a133c288c85b51f5f5a60e0fe7913b20f9fd4a9bf8104561e23e33479dada46" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.835967 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc160003-2b7c-4772-a0f3-7b431c501fb8-combined-ca-bundle\") pod \"swift-proxy-59d5c55c95-qdh4t\" (UID: \"cc160003-2b7c-4772-a0f3-7b431c501fb8\") " pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.839919 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a133c288c85b51f5f5a60e0fe7913b20f9fd4a9bf8104561e23e33479dada46"} err="failed to get container status \"9a133c288c85b51f5f5a60e0fe7913b20f9fd4a9bf8104561e23e33479dada46\": rpc error: code = NotFound desc = could not find container \"9a133c288c85b51f5f5a60e0fe7913b20f9fd4a9bf8104561e23e33479dada46\": container with ID starting with 9a133c288c85b51f5f5a60e0fe7913b20f9fd4a9bf8104561e23e33479dada46 not found: ID does not exist" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.839976 4998 scope.go:117] "RemoveContainer" containerID="47ad5ff218d1db4e4a839220fd027f05fd704ae47e711f9827767480718bfd25" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.840789 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc160003-2b7c-4772-a0f3-7b431c501fb8-config-data\") pod \"swift-proxy-59d5c55c95-qdh4t\" (UID: \"cc160003-2b7c-4772-a0f3-7b431c501fb8\") " pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.841905 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cc160003-2b7c-4772-a0f3-7b431c501fb8-etc-swift\") pod \"swift-proxy-59d5c55c95-qdh4t\" (UID: \"cc160003-2b7c-4772-a0f3-7b431c501fb8\") " pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.847512 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xm7l\" (UniqueName: \"kubernetes.io/projected/cc160003-2b7c-4772-a0f3-7b431c501fb8-kube-api-access-9xm7l\") pod \"swift-proxy-59d5c55c95-qdh4t\" (UID: \"cc160003-2b7c-4772-a0f3-7b431c501fb8\") " pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.851851 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc160003-2b7c-4772-a0f3-7b431c501fb8-internal-tls-certs\") pod \"swift-proxy-59d5c55c95-qdh4t\" (UID: \"cc160003-2b7c-4772-a0f3-7b431c501fb8\") " pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.852505 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc160003-2b7c-4772-a0f3-7b431c501fb8-public-tls-certs\") pod \"swift-proxy-59d5c55c95-qdh4t\" (UID: \"cc160003-2b7c-4772-a0f3-7b431c501fb8\") " pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.855529 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47ad5ff218d1db4e4a839220fd027f05fd704ae47e711f9827767480718bfd25"} err="failed to get container status \"47ad5ff218d1db4e4a839220fd027f05fd704ae47e711f9827767480718bfd25\": rpc error: code = NotFound desc = could not find container \"47ad5ff218d1db4e4a839220fd027f05fd704ae47e711f9827767480718bfd25\": container with ID starting with 47ad5ff218d1db4e4a839220fd027f05fd704ae47e711f9827767480718bfd25 not found: ID does not exist" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.882653 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.910862 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.922747 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.924438 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.932165 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.942513 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 16:25:33 crc kubenswrapper[4998]: I1203 16:25:33.994809 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:34 crc kubenswrapper[4998]: I1203 16:25:34.028813 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4642ed6c-0202-4a7d-b50d-9c90247cfdb9-config-data\") pod \"cinder-scheduler-0\" (UID: \"4642ed6c-0202-4a7d-b50d-9c90247cfdb9\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:34 crc kubenswrapper[4998]: I1203 16:25:34.028937 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4642ed6c-0202-4a7d-b50d-9c90247cfdb9-scripts\") pod \"cinder-scheduler-0\" (UID: \"4642ed6c-0202-4a7d-b50d-9c90247cfdb9\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:34 crc kubenswrapper[4998]: I1203 16:25:34.029042 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4642ed6c-0202-4a7d-b50d-9c90247cfdb9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4642ed6c-0202-4a7d-b50d-9c90247cfdb9\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:34 crc kubenswrapper[4998]: I1203 16:25:34.029072 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4642ed6c-0202-4a7d-b50d-9c90247cfdb9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4642ed6c-0202-4a7d-b50d-9c90247cfdb9\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:34 crc kubenswrapper[4998]: I1203 16:25:34.029156 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntbmv\" (UniqueName: \"kubernetes.io/projected/4642ed6c-0202-4a7d-b50d-9c90247cfdb9-kube-api-access-ntbmv\") pod \"cinder-scheduler-0\" (UID: \"4642ed6c-0202-4a7d-b50d-9c90247cfdb9\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:34 crc kubenswrapper[4998]: I1203 16:25:34.029209 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4642ed6c-0202-4a7d-b50d-9c90247cfdb9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4642ed6c-0202-4a7d-b50d-9c90247cfdb9\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:34 crc kubenswrapper[4998]: I1203 16:25:34.132285 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4642ed6c-0202-4a7d-b50d-9c90247cfdb9-scripts\") pod \"cinder-scheduler-0\" (UID: \"4642ed6c-0202-4a7d-b50d-9c90247cfdb9\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:34 crc kubenswrapper[4998]: I1203 16:25:34.132540 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4642ed6c-0202-4a7d-b50d-9c90247cfdb9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4642ed6c-0202-4a7d-b50d-9c90247cfdb9\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:34 crc kubenswrapper[4998]: I1203 16:25:34.132557 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4642ed6c-0202-4a7d-b50d-9c90247cfdb9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4642ed6c-0202-4a7d-b50d-9c90247cfdb9\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:34 crc kubenswrapper[4998]: I1203 16:25:34.132614 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntbmv\" (UniqueName: \"kubernetes.io/projected/4642ed6c-0202-4a7d-b50d-9c90247cfdb9-kube-api-access-ntbmv\") pod \"cinder-scheduler-0\" (UID: \"4642ed6c-0202-4a7d-b50d-9c90247cfdb9\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:34 crc kubenswrapper[4998]: I1203 16:25:34.132648 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4642ed6c-0202-4a7d-b50d-9c90247cfdb9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4642ed6c-0202-4a7d-b50d-9c90247cfdb9\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:34 crc kubenswrapper[4998]: I1203 16:25:34.132703 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4642ed6c-0202-4a7d-b50d-9c90247cfdb9-config-data\") pod \"cinder-scheduler-0\" (UID: \"4642ed6c-0202-4a7d-b50d-9c90247cfdb9\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:34 crc kubenswrapper[4998]: I1203 16:25:34.138255 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4642ed6c-0202-4a7d-b50d-9c90247cfdb9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4642ed6c-0202-4a7d-b50d-9c90247cfdb9\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:34 crc kubenswrapper[4998]: I1203 16:25:34.138744 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4642ed6c-0202-4a7d-b50d-9c90247cfdb9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4642ed6c-0202-4a7d-b50d-9c90247cfdb9\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:34 crc kubenswrapper[4998]: I1203 16:25:34.140336 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4642ed6c-0202-4a7d-b50d-9c90247cfdb9-config-data\") pod \"cinder-scheduler-0\" (UID: \"4642ed6c-0202-4a7d-b50d-9c90247cfdb9\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:34 crc kubenswrapper[4998]: I1203 16:25:34.142075 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4642ed6c-0202-4a7d-b50d-9c90247cfdb9-scripts\") pod \"cinder-scheduler-0\" (UID: \"4642ed6c-0202-4a7d-b50d-9c90247cfdb9\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:34 crc kubenswrapper[4998]: I1203 16:25:34.148305 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4642ed6c-0202-4a7d-b50d-9c90247cfdb9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4642ed6c-0202-4a7d-b50d-9c90247cfdb9\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:34 crc kubenswrapper[4998]: I1203 16:25:34.157712 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntbmv\" (UniqueName: \"kubernetes.io/projected/4642ed6c-0202-4a7d-b50d-9c90247cfdb9-kube-api-access-ntbmv\") pod \"cinder-scheduler-0\" (UID: \"4642ed6c-0202-4a7d-b50d-9c90247cfdb9\") " pod="openstack/cinder-scheduler-0" Dec 03 16:25:34 crc kubenswrapper[4998]: I1203 16:25:34.260058 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 16:25:34 crc kubenswrapper[4998]: I1203 16:25:34.488024 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77c59f659f-jkw4x" event={"ID":"3aa7afc0-1cf1-4e6a-ae4e-98b13411094c","Type":"ContainerStarted","Data":"cf4cedd382e202bbba290fcad0e59adbc59456ada7c157f59715d3be02040f95"} Dec 03 16:25:34 crc kubenswrapper[4998]: I1203 16:25:34.498857 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-77c59f659f-jkw4x" Dec 03 16:25:34 crc kubenswrapper[4998]: I1203 16:25:34.504612 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f32474ed-1700-42c9-b87d-ffc1f8445ea6","Type":"ContainerStarted","Data":"97dd6453d32f7151528140a97eace904d6bd6eb130c780341f43e4e246d4e9a3"} Dec 03 16:25:34 crc kubenswrapper[4998]: I1203 16:25:34.567639 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-77c59f659f-jkw4x" podStartSLOduration=4.567622169 podStartE2EDuration="4.567622169s" podCreationTimestamp="2025-12-03 16:25:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:25:34.530096596 +0000 UTC m=+1313.141796819" watchObservedRunningTime="2025-12-03 16:25:34.567622169 +0000 UTC m=+1313.179322382" Dec 03 16:25:34 crc kubenswrapper[4998]: I1203 16:25:34.774176 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-59d5c55c95-qdh4t"] Dec 03 16:25:34 crc kubenswrapper[4998]: I1203 16:25:34.983013 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 16:25:35 crc kubenswrapper[4998]: I1203 16:25:35.527301 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1ee49423-2774-4d51-ae4f-8cea1775efef","Type":"ContainerStarted","Data":"fee3c625913536f8b096ea869d6f649f893ecdad4891bd241dfe418afdb24fb2"} Dec 03 16:25:35 crc kubenswrapper[4998]: I1203 16:25:35.536498 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f32474ed-1700-42c9-b87d-ffc1f8445ea6","Type":"ContainerStarted","Data":"800894309e99b901dd8864c7e28ee2fe0b7559f645f17ca3b56219578d4937e4"} Dec 03 16:25:35 crc kubenswrapper[4998]: I1203 16:25:35.546251 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4642ed6c-0202-4a7d-b50d-9c90247cfdb9","Type":"ContainerStarted","Data":"c3f6bdefe6b5233376204defe1cfad62c6b7463e3b2d1500c0d79688d6a27206"} Dec 03 16:25:35 crc kubenswrapper[4998]: I1203 16:25:35.557554 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-59d5c55c95-qdh4t" event={"ID":"cc160003-2b7c-4772-a0f3-7b431c501fb8","Type":"ContainerStarted","Data":"5ed530e09a0786d9e770494933b704e2dc15dee22f96e9c3e51d148a0af46047"} Dec 03 16:25:35 crc kubenswrapper[4998]: I1203 16:25:35.557591 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-59d5c55c95-qdh4t" event={"ID":"cc160003-2b7c-4772-a0f3-7b431c501fb8","Type":"ContainerStarted","Data":"95b4b853eda3a3cabc43c51af3b9dbc2bd739cb3f4d79bf12e58aeaa1e761155"} Dec 03 16:25:35 crc kubenswrapper[4998]: I1203 16:25:35.565607 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.176:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 16:25:35 crc kubenswrapper[4998]: I1203 16:25:35.629121 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 16:25:35 crc kubenswrapper[4998]: I1203 16:25:35.714024 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6872e78-e3b5-4024-9de0-24c6a3cfe345" path="/var/lib/kubelet/pods/e6872e78-e3b5-4024-9de0-24c6a3cfe345/volumes" Dec 03 16:25:35 crc kubenswrapper[4998]: I1203 16:25:35.714897 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 16:25:35 crc kubenswrapper[4998]: I1203 16:25:35.912884 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:25:35 crc kubenswrapper[4998]: I1203 16:25:35.913152 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="855663b4-d4de-4ca0-811c-44d4def02706" containerName="ceilometer-central-agent" containerID="cri-o://bd642294c5a767432620ab6245505b9509dcd42d5a1e0c7143d68ee5baa5b382" gracePeriod=30 Dec 03 16:25:35 crc kubenswrapper[4998]: I1203 16:25:35.916164 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="855663b4-d4de-4ca0-811c-44d4def02706" containerName="proxy-httpd" containerID="cri-o://f75e25cdac20077210bd49dc9952da9b52dc8acf1d5253701e494daa9d79e244" gracePeriod=30 Dec 03 16:25:35 crc kubenswrapper[4998]: I1203 16:25:35.916327 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="855663b4-d4de-4ca0-811c-44d4def02706" containerName="ceilometer-notification-agent" containerID="cri-o://2c63eb0b5ccdeff1d9234d2c6cc21f2aab83a23b017acd58cc9a301f5ca959c6" gracePeriod=30 Dec 03 16:25:35 crc kubenswrapper[4998]: I1203 16:25:35.916368 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="855663b4-d4de-4ca0-811c-44d4def02706" containerName="sg-core" containerID="cri-o://d467c7642ab7b67ec0bb6e94bf7954b5748a94ef142fb733a8fe87b353b6838d" gracePeriod=30 Dec 03 16:25:35 crc kubenswrapper[4998]: I1203 16:25:35.937316 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="855663b4-d4de-4ca0-811c-44d4def02706" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.173:3000/\": EOF" Dec 03 16:25:36 crc kubenswrapper[4998]: I1203 16:25:36.586506 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-59d5c55c95-qdh4t" event={"ID":"cc160003-2b7c-4772-a0f3-7b431c501fb8","Type":"ContainerStarted","Data":"97472348a0814fc66073b649e4df69ebae1d4d0db221620fedb59bbc6921c699"} Dec 03 16:25:36 crc kubenswrapper[4998]: I1203 16:25:36.587181 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:36 crc kubenswrapper[4998]: I1203 16:25:36.587262 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:36 crc kubenswrapper[4998]: I1203 16:25:36.593864 4998 generic.go:334] "Generic (PLEG): container finished" podID="855663b4-d4de-4ca0-811c-44d4def02706" containerID="f75e25cdac20077210bd49dc9952da9b52dc8acf1d5253701e494daa9d79e244" exitCode=0 Dec 03 16:25:36 crc kubenswrapper[4998]: I1203 16:25:36.593907 4998 generic.go:334] "Generic (PLEG): container finished" podID="855663b4-d4de-4ca0-811c-44d4def02706" containerID="d467c7642ab7b67ec0bb6e94bf7954b5748a94ef142fb733a8fe87b353b6838d" exitCode=2 Dec 03 16:25:36 crc kubenswrapper[4998]: I1203 16:25:36.593971 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"855663b4-d4de-4ca0-811c-44d4def02706","Type":"ContainerDied","Data":"f75e25cdac20077210bd49dc9952da9b52dc8acf1d5253701e494daa9d79e244"} Dec 03 16:25:36 crc kubenswrapper[4998]: I1203 16:25:36.593998 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"855663b4-d4de-4ca0-811c-44d4def02706","Type":"ContainerDied","Data":"d467c7642ab7b67ec0bb6e94bf7954b5748a94ef142fb733a8fe87b353b6838d"} Dec 03 16:25:36 crc kubenswrapper[4998]: I1203 16:25:36.597458 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="1ee49423-2774-4d51-ae4f-8cea1775efef" containerName="glance-log" containerID="cri-o://fee3c625913536f8b096ea869d6f649f893ecdad4891bd241dfe418afdb24fb2" gracePeriod=30 Dec 03 16:25:36 crc kubenswrapper[4998]: I1203 16:25:36.597666 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1ee49423-2774-4d51-ae4f-8cea1775efef","Type":"ContainerStarted","Data":"3edbe408fc0f5fdbc06f520e34dfd518b9ba486182e872e816e6c7ee46e55a09"} Dec 03 16:25:36 crc kubenswrapper[4998]: I1203 16:25:36.597717 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="1ee49423-2774-4d51-ae4f-8cea1775efef" containerName="glance-httpd" containerID="cri-o://3edbe408fc0f5fdbc06f520e34dfd518b9ba486182e872e816e6c7ee46e55a09" gracePeriod=30 Dec 03 16:25:36 crc kubenswrapper[4998]: I1203 16:25:36.606866 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f32474ed-1700-42c9-b87d-ffc1f8445ea6","Type":"ContainerStarted","Data":"7f1a10aa4caf660e44bbd064ed15a89696116af59f920f9237b9113d9d846b37"} Dec 03 16:25:36 crc kubenswrapper[4998]: I1203 16:25:36.607094 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f32474ed-1700-42c9-b87d-ffc1f8445ea6" containerName="glance-log" containerID="cri-o://800894309e99b901dd8864c7e28ee2fe0b7559f645f17ca3b56219578d4937e4" gracePeriod=30 Dec 03 16:25:36 crc kubenswrapper[4998]: I1203 16:25:36.607405 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f32474ed-1700-42c9-b87d-ffc1f8445ea6" containerName="glance-httpd" containerID="cri-o://7f1a10aa4caf660e44bbd064ed15a89696116af59f920f9237b9113d9d846b37" gracePeriod=30 Dec 03 16:25:36 crc kubenswrapper[4998]: I1203 16:25:36.615124 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4642ed6c-0202-4a7d-b50d-9c90247cfdb9","Type":"ContainerStarted","Data":"b03cce4a1fc1518046031b051cc9aa48497b22c49bb1a292ced4d13309bfb5b6"} Dec 03 16:25:36 crc kubenswrapper[4998]: I1203 16:25:36.623592 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-59d5c55c95-qdh4t" podStartSLOduration=3.6235738 podStartE2EDuration="3.6235738s" podCreationTimestamp="2025-12-03 16:25:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:25:36.615577865 +0000 UTC m=+1315.227278118" watchObservedRunningTime="2025-12-03 16:25:36.6235738 +0000 UTC m=+1315.235274023" Dec 03 16:25:36 crc kubenswrapper[4998]: I1203 16:25:36.642653 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.642637094 podStartE2EDuration="6.642637094s" podCreationTimestamp="2025-12-03 16:25:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:25:36.641006484 +0000 UTC m=+1315.252706707" watchObservedRunningTime="2025-12-03 16:25:36.642637094 +0000 UTC m=+1315.254337317" Dec 03 16:25:36 crc kubenswrapper[4998]: I1203 16:25:36.675338 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.675319069 podStartE2EDuration="6.675319069s" podCreationTimestamp="2025-12-03 16:25:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:25:36.666429113 +0000 UTC m=+1315.278129356" watchObservedRunningTime="2025-12-03 16:25:36.675319069 +0000 UTC m=+1315.287019302" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.389126 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.468978 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1ee49423-2774-4d51-ae4f-8cea1775efef-httpd-run\") pod \"1ee49423-2774-4d51-ae4f-8cea1775efef\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.469049 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqf6x\" (UniqueName: \"kubernetes.io/projected/1ee49423-2774-4d51-ae4f-8cea1775efef-kube-api-access-sqf6x\") pod \"1ee49423-2774-4d51-ae4f-8cea1775efef\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.469115 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ee49423-2774-4d51-ae4f-8cea1775efef-combined-ca-bundle\") pod \"1ee49423-2774-4d51-ae4f-8cea1775efef\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.469138 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ee49423-2774-4d51-ae4f-8cea1775efef-config-data\") pod \"1ee49423-2774-4d51-ae4f-8cea1775efef\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.469153 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"1ee49423-2774-4d51-ae4f-8cea1775efef\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.469214 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ee49423-2774-4d51-ae4f-8cea1775efef-scripts\") pod \"1ee49423-2774-4d51-ae4f-8cea1775efef\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.469356 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ee49423-2774-4d51-ae4f-8cea1775efef-logs\") pod \"1ee49423-2774-4d51-ae4f-8cea1775efef\" (UID: \"1ee49423-2774-4d51-ae4f-8cea1775efef\") " Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.470374 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ee49423-2774-4d51-ae4f-8cea1775efef-logs" (OuterVolumeSpecName: "logs") pod "1ee49423-2774-4d51-ae4f-8cea1775efef" (UID: "1ee49423-2774-4d51-ae4f-8cea1775efef"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.471903 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ee49423-2774-4d51-ae4f-8cea1775efef-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1ee49423-2774-4d51-ae4f-8cea1775efef" (UID: "1ee49423-2774-4d51-ae4f-8cea1775efef"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.478911 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ee49423-2774-4d51-ae4f-8cea1775efef-kube-api-access-sqf6x" (OuterVolumeSpecName: "kube-api-access-sqf6x") pod "1ee49423-2774-4d51-ae4f-8cea1775efef" (UID: "1ee49423-2774-4d51-ae4f-8cea1775efef"). InnerVolumeSpecName "kube-api-access-sqf6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.479418 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "1ee49423-2774-4d51-ae4f-8cea1775efef" (UID: "1ee49423-2774-4d51-ae4f-8cea1775efef"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.482698 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ee49423-2774-4d51-ae4f-8cea1775efef-scripts" (OuterVolumeSpecName: "scripts") pod "1ee49423-2774-4d51-ae4f-8cea1775efef" (UID: "1ee49423-2774-4d51-ae4f-8cea1775efef"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.506648 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ee49423-2774-4d51-ae4f-8cea1775efef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1ee49423-2774-4d51-ae4f-8cea1775efef" (UID: "1ee49423-2774-4d51-ae4f-8cea1775efef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.531928 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ee49423-2774-4d51-ae4f-8cea1775efef-config-data" (OuterVolumeSpecName: "config-data") pod "1ee49423-2774-4d51-ae4f-8cea1775efef" (UID: "1ee49423-2774-4d51-ae4f-8cea1775efef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.572031 4998 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ee49423-2774-4d51-ae4f-8cea1775efef-logs\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.572083 4998 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1ee49423-2774-4d51-ae4f-8cea1775efef-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.572099 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqf6x\" (UniqueName: \"kubernetes.io/projected/1ee49423-2774-4d51-ae4f-8cea1775efef-kube-api-access-sqf6x\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.572111 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ee49423-2774-4d51-ae4f-8cea1775efef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.572124 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ee49423-2774-4d51-ae4f-8cea1775efef-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.572163 4998 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.572176 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ee49423-2774-4d51-ae4f-8cea1775efef-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.605048 4998 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.615009 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.673772 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnpp7\" (UniqueName: \"kubernetes.io/projected/f32474ed-1700-42c9-b87d-ffc1f8445ea6-kube-api-access-lnpp7\") pod \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.673824 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f32474ed-1700-42c9-b87d-ffc1f8445ea6-scripts\") pod \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.673912 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f32474ed-1700-42c9-b87d-ffc1f8445ea6-httpd-run\") pod \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.673946 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f32474ed-1700-42c9-b87d-ffc1f8445ea6-combined-ca-bundle\") pod \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.673973 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f32474ed-1700-42c9-b87d-ffc1f8445ea6-config-data\") pod \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.674075 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f32474ed-1700-42c9-b87d-ffc1f8445ea6-logs\") pod \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.674113 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\" (UID: \"f32474ed-1700-42c9-b87d-ffc1f8445ea6\") " Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.674474 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f32474ed-1700-42c9-b87d-ffc1f8445ea6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f32474ed-1700-42c9-b87d-ffc1f8445ea6" (UID: "f32474ed-1700-42c9-b87d-ffc1f8445ea6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.674592 4998 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.674629 4998 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f32474ed-1700-42c9-b87d-ffc1f8445ea6-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.675811 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f32474ed-1700-42c9-b87d-ffc1f8445ea6-logs" (OuterVolumeSpecName: "logs") pod "f32474ed-1700-42c9-b87d-ffc1f8445ea6" (UID: "f32474ed-1700-42c9-b87d-ffc1f8445ea6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.683677 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f32474ed-1700-42c9-b87d-ffc1f8445ea6-kube-api-access-lnpp7" (OuterVolumeSpecName: "kube-api-access-lnpp7") pod "f32474ed-1700-42c9-b87d-ffc1f8445ea6" (UID: "f32474ed-1700-42c9-b87d-ffc1f8445ea6"). InnerVolumeSpecName "kube-api-access-lnpp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.684377 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "f32474ed-1700-42c9-b87d-ffc1f8445ea6" (UID: "f32474ed-1700-42c9-b87d-ffc1f8445ea6"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.685859 4998 generic.go:334] "Generic (PLEG): container finished" podID="a009aeee-c849-4c00-b41b-3cf5e36a75bb" containerID="b657ed22e9096f43b7242cbed330ab7de4b3e3c8d87141fb5d8555f0560c2457" exitCode=1 Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.691668 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f32474ed-1700-42c9-b87d-ffc1f8445ea6-scripts" (OuterVolumeSpecName: "scripts") pod "f32474ed-1700-42c9-b87d-ffc1f8445ea6" (UID: "f32474ed-1700-42c9-b87d-ffc1f8445ea6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.705023 4998 generic.go:334] "Generic (PLEG): container finished" podID="855663b4-d4de-4ca0-811c-44d4def02706" containerID="bd642294c5a767432620ab6245505b9509dcd42d5a1e0c7143d68ee5baa5b382" exitCode=0 Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.709554 4998 generic.go:334] "Generic (PLEG): container finished" podID="1ee49423-2774-4d51-ae4f-8cea1775efef" containerID="3edbe408fc0f5fdbc06f520e34dfd518b9ba486182e872e816e6c7ee46e55a09" exitCode=143 Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.709573 4998 generic.go:334] "Generic (PLEG): container finished" podID="1ee49423-2774-4d51-ae4f-8cea1775efef" containerID="fee3c625913536f8b096ea869d6f649f893ecdad4891bd241dfe418afdb24fb2" exitCode=143 Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.709639 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.711747 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f32474ed-1700-42c9-b87d-ffc1f8445ea6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f32474ed-1700-42c9-b87d-ffc1f8445ea6" (UID: "f32474ed-1700-42c9-b87d-ffc1f8445ea6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.718470 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.718456594 podStartE2EDuration="4.718456594s" podCreationTimestamp="2025-12-03 16:25:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:25:37.717174013 +0000 UTC m=+1316.328874246" watchObservedRunningTime="2025-12-03 16:25:37.718456594 +0000 UTC m=+1316.330156817" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.721065 4998 generic.go:334] "Generic (PLEG): container finished" podID="f32474ed-1700-42c9-b87d-ffc1f8445ea6" containerID="7f1a10aa4caf660e44bbd064ed15a89696116af59f920f9237b9113d9d846b37" exitCode=143 Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.721093 4998 generic.go:334] "Generic (PLEG): container finished" podID="f32474ed-1700-42c9-b87d-ffc1f8445ea6" containerID="800894309e99b901dd8864c7e28ee2fe0b7559f645f17ca3b56219578d4937e4" exitCode=143 Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.721524 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.723301 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"a009aeee-c849-4c00-b41b-3cf5e36a75bb","Type":"ContainerDied","Data":"b657ed22e9096f43b7242cbed330ab7de4b3e3c8d87141fb5d8555f0560c2457"} Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.723418 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4642ed6c-0202-4a7d-b50d-9c90247cfdb9","Type":"ContainerStarted","Data":"b7def468d2dcaf31ac820d415ec0b82d7a919930f9991c6483d901d9ad2fa3a4"} Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.723487 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"855663b4-d4de-4ca0-811c-44d4def02706","Type":"ContainerDied","Data":"bd642294c5a767432620ab6245505b9509dcd42d5a1e0c7143d68ee5baa5b382"} Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.723551 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1ee49423-2774-4d51-ae4f-8cea1775efef","Type":"ContainerDied","Data":"3edbe408fc0f5fdbc06f520e34dfd518b9ba486182e872e816e6c7ee46e55a09"} Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.723636 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1ee49423-2774-4d51-ae4f-8cea1775efef","Type":"ContainerDied","Data":"fee3c625913536f8b096ea869d6f649f893ecdad4891bd241dfe418afdb24fb2"} Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.723705 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1ee49423-2774-4d51-ae4f-8cea1775efef","Type":"ContainerDied","Data":"97bbe409920beffc63a9ade541117bff7a7e66ee697bbf668c455419561bfdc4"} Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.723780 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f32474ed-1700-42c9-b87d-ffc1f8445ea6","Type":"ContainerDied","Data":"7f1a10aa4caf660e44bbd064ed15a89696116af59f920f9237b9113d9d846b37"} Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.723869 4998 scope.go:117] "RemoveContainer" containerID="616cb19e04d79255ec317f93c152a8d4bf88d542e297bf72a69bea31d5e86255" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.723923 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f32474ed-1700-42c9-b87d-ffc1f8445ea6","Type":"ContainerDied","Data":"800894309e99b901dd8864c7e28ee2fe0b7559f645f17ca3b56219578d4937e4"} Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.724129 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f32474ed-1700-42c9-b87d-ffc1f8445ea6","Type":"ContainerDied","Data":"97dd6453d32f7151528140a97eace904d6bd6eb130c780341f43e4e246d4e9a3"} Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.724632 4998 scope.go:117] "RemoveContainer" containerID="b657ed22e9096f43b7242cbed330ab7de4b3e3c8d87141fb5d8555f0560c2457" Dec 03 16:25:37 crc kubenswrapper[4998]: E1203 16:25:37.724868 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(a009aeee-c849-4c00-b41b-3cf5e36a75bb)\"" pod="openstack/watcher-decision-engine-0" podUID="a009aeee-c849-4c00-b41b-3cf5e36a75bb" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.765962 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f32474ed-1700-42c9-b87d-ffc1f8445ea6-config-data" (OuterVolumeSpecName: "config-data") pod "f32474ed-1700-42c9-b87d-ffc1f8445ea6" (UID: "f32474ed-1700-42c9-b87d-ffc1f8445ea6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.782293 4998 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f32474ed-1700-42c9-b87d-ffc1f8445ea6-logs\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.782332 4998 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.782343 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnpp7\" (UniqueName: \"kubernetes.io/projected/f32474ed-1700-42c9-b87d-ffc1f8445ea6-kube-api-access-lnpp7\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.782353 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f32474ed-1700-42c9-b87d-ffc1f8445ea6-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.782362 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f32474ed-1700-42c9-b87d-ffc1f8445ea6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.782370 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f32474ed-1700-42c9-b87d-ffc1f8445ea6-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.808988 4998 scope.go:117] "RemoveContainer" containerID="3edbe408fc0f5fdbc06f520e34dfd518b9ba486182e872e816e6c7ee46e55a09" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.811912 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.827243 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.836509 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 16:25:37 crc kubenswrapper[4998]: E1203 16:25:37.836961 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ee49423-2774-4d51-ae4f-8cea1775efef" containerName="glance-httpd" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.836985 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ee49423-2774-4d51-ae4f-8cea1775efef" containerName="glance-httpd" Dec 03 16:25:37 crc kubenswrapper[4998]: E1203 16:25:37.836997 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f32474ed-1700-42c9-b87d-ffc1f8445ea6" containerName="glance-log" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.837004 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="f32474ed-1700-42c9-b87d-ffc1f8445ea6" containerName="glance-log" Dec 03 16:25:37 crc kubenswrapper[4998]: E1203 16:25:37.837013 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ee49423-2774-4d51-ae4f-8cea1775efef" containerName="glance-log" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.837019 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ee49423-2774-4d51-ae4f-8cea1775efef" containerName="glance-log" Dec 03 16:25:37 crc kubenswrapper[4998]: E1203 16:25:37.837031 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f32474ed-1700-42c9-b87d-ffc1f8445ea6" containerName="glance-httpd" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.837038 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="f32474ed-1700-42c9-b87d-ffc1f8445ea6" containerName="glance-httpd" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.837240 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="f32474ed-1700-42c9-b87d-ffc1f8445ea6" containerName="glance-log" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.837263 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ee49423-2774-4d51-ae4f-8cea1775efef" containerName="glance-log" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.837273 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="f32474ed-1700-42c9-b87d-ffc1f8445ea6" containerName="glance-httpd" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.837279 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ee49423-2774-4d51-ae4f-8cea1775efef" containerName="glance-httpd" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.838360 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.840320 4998 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.843199 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.843299 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.843409 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.843708 4998 scope.go:117] "RemoveContainer" containerID="fee3c625913536f8b096ea869d6f649f893ecdad4891bd241dfe418afdb24fb2" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.884475 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.884556 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.884589 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-logs\") pod \"glance-default-external-api-0\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.884607 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.884628 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-scripts\") pod \"glance-default-external-api-0\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.884660 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-config-data\") pod \"glance-default-external-api-0\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.884686 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z96p\" (UniqueName: \"kubernetes.io/projected/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-kube-api-access-8z96p\") pod \"glance-default-external-api-0\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.884717 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.884832 4998 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.973943 4998 scope.go:117] "RemoveContainer" containerID="3edbe408fc0f5fdbc06f520e34dfd518b9ba486182e872e816e6c7ee46e55a09" Dec 03 16:25:37 crc kubenswrapper[4998]: E1203 16:25:37.978773 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3edbe408fc0f5fdbc06f520e34dfd518b9ba486182e872e816e6c7ee46e55a09\": container with ID starting with 3edbe408fc0f5fdbc06f520e34dfd518b9ba486182e872e816e6c7ee46e55a09 not found: ID does not exist" containerID="3edbe408fc0f5fdbc06f520e34dfd518b9ba486182e872e816e6c7ee46e55a09" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.978805 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3edbe408fc0f5fdbc06f520e34dfd518b9ba486182e872e816e6c7ee46e55a09"} err="failed to get container status \"3edbe408fc0f5fdbc06f520e34dfd518b9ba486182e872e816e6c7ee46e55a09\": rpc error: code = NotFound desc = could not find container \"3edbe408fc0f5fdbc06f520e34dfd518b9ba486182e872e816e6c7ee46e55a09\": container with ID starting with 3edbe408fc0f5fdbc06f520e34dfd518b9ba486182e872e816e6c7ee46e55a09 not found: ID does not exist" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.978826 4998 scope.go:117] "RemoveContainer" containerID="fee3c625913536f8b096ea869d6f649f893ecdad4891bd241dfe418afdb24fb2" Dec 03 16:25:37 crc kubenswrapper[4998]: E1203 16:25:37.983160 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fee3c625913536f8b096ea869d6f649f893ecdad4891bd241dfe418afdb24fb2\": container with ID starting with fee3c625913536f8b096ea869d6f649f893ecdad4891bd241dfe418afdb24fb2 not found: ID does not exist" containerID="fee3c625913536f8b096ea869d6f649f893ecdad4891bd241dfe418afdb24fb2" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.983196 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fee3c625913536f8b096ea869d6f649f893ecdad4891bd241dfe418afdb24fb2"} err="failed to get container status \"fee3c625913536f8b096ea869d6f649f893ecdad4891bd241dfe418afdb24fb2\": rpc error: code = NotFound desc = could not find container \"fee3c625913536f8b096ea869d6f649f893ecdad4891bd241dfe418afdb24fb2\": container with ID starting with fee3c625913536f8b096ea869d6f649f893ecdad4891bd241dfe418afdb24fb2 not found: ID does not exist" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.983221 4998 scope.go:117] "RemoveContainer" containerID="3edbe408fc0f5fdbc06f520e34dfd518b9ba486182e872e816e6c7ee46e55a09" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.986855 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3edbe408fc0f5fdbc06f520e34dfd518b9ba486182e872e816e6c7ee46e55a09"} err="failed to get container status \"3edbe408fc0f5fdbc06f520e34dfd518b9ba486182e872e816e6c7ee46e55a09\": rpc error: code = NotFound desc = could not find container \"3edbe408fc0f5fdbc06f520e34dfd518b9ba486182e872e816e6c7ee46e55a09\": container with ID starting with 3edbe408fc0f5fdbc06f520e34dfd518b9ba486182e872e816e6c7ee46e55a09 not found: ID does not exist" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.986890 4998 scope.go:117] "RemoveContainer" containerID="fee3c625913536f8b096ea869d6f649f893ecdad4891bd241dfe418afdb24fb2" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.988238 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fee3c625913536f8b096ea869d6f649f893ecdad4891bd241dfe418afdb24fb2"} err="failed to get container status \"fee3c625913536f8b096ea869d6f649f893ecdad4891bd241dfe418afdb24fb2\": rpc error: code = NotFound desc = could not find container \"fee3c625913536f8b096ea869d6f649f893ecdad4891bd241dfe418afdb24fb2\": container with ID starting with fee3c625913536f8b096ea869d6f649f893ecdad4891bd241dfe418afdb24fb2 not found: ID does not exist" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.988260 4998 scope.go:117] "RemoveContainer" containerID="7f1a10aa4caf660e44bbd064ed15a89696116af59f920f9237b9113d9d846b37" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.988389 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.988477 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.988526 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.988553 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-logs\") pod \"glance-default-external-api-0\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.988570 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.988589 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-scripts\") pod \"glance-default-external-api-0\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.988616 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-config-data\") pod \"glance-default-external-api-0\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.988643 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z96p\" (UniqueName: \"kubernetes.io/projected/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-kube-api-access-8z96p\") pod \"glance-default-external-api-0\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.989102 4998 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.989814 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-logs\") pod \"glance-default-external-api-0\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.993583 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:37 crc kubenswrapper[4998]: I1203 16:25:37.994531 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.000522 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-scripts\") pod \"glance-default-external-api-0\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.003060 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-config-data\") pod \"glance-default-external-api-0\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.004447 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.016377 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z96p\" (UniqueName: \"kubernetes.io/projected/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-kube-api-access-8z96p\") pod \"glance-default-external-api-0\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.042972 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " pod="openstack/glance-default-external-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.052528 4998 scope.go:117] "RemoveContainer" containerID="800894309e99b901dd8864c7e28ee2fe0b7559f645f17ca3b56219578d4937e4" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.089996 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.107286 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.111622 4998 scope.go:117] "RemoveContainer" containerID="7f1a10aa4caf660e44bbd064ed15a89696116af59f920f9237b9113d9d846b37" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.114444 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 16:25:38 crc kubenswrapper[4998]: E1203 16:25:38.114871 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f1a10aa4caf660e44bbd064ed15a89696116af59f920f9237b9113d9d846b37\": container with ID starting with 7f1a10aa4caf660e44bbd064ed15a89696116af59f920f9237b9113d9d846b37 not found: ID does not exist" containerID="7f1a10aa4caf660e44bbd064ed15a89696116af59f920f9237b9113d9d846b37" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.114912 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f1a10aa4caf660e44bbd064ed15a89696116af59f920f9237b9113d9d846b37"} err="failed to get container status \"7f1a10aa4caf660e44bbd064ed15a89696116af59f920f9237b9113d9d846b37\": rpc error: code = NotFound desc = could not find container \"7f1a10aa4caf660e44bbd064ed15a89696116af59f920f9237b9113d9d846b37\": container with ID starting with 7f1a10aa4caf660e44bbd064ed15a89696116af59f920f9237b9113d9d846b37 not found: ID does not exist" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.114941 4998 scope.go:117] "RemoveContainer" containerID="800894309e99b901dd8864c7e28ee2fe0b7559f645f17ca3b56219578d4937e4" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.120074 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: E1203 16:25:38.122229 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"800894309e99b901dd8864c7e28ee2fe0b7559f645f17ca3b56219578d4937e4\": container with ID starting with 800894309e99b901dd8864c7e28ee2fe0b7559f645f17ca3b56219578d4937e4 not found: ID does not exist" containerID="800894309e99b901dd8864c7e28ee2fe0b7559f645f17ca3b56219578d4937e4" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.122263 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"800894309e99b901dd8864c7e28ee2fe0b7559f645f17ca3b56219578d4937e4"} err="failed to get container status \"800894309e99b901dd8864c7e28ee2fe0b7559f645f17ca3b56219578d4937e4\": rpc error: code = NotFound desc = could not find container \"800894309e99b901dd8864c7e28ee2fe0b7559f645f17ca3b56219578d4937e4\": container with ID starting with 800894309e99b901dd8864c7e28ee2fe0b7559f645f17ca3b56219578d4937e4 not found: ID does not exist" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.122288 4998 scope.go:117] "RemoveContainer" containerID="7f1a10aa4caf660e44bbd064ed15a89696116af59f920f9237b9113d9d846b37" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.124139 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.124342 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.128352 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f1a10aa4caf660e44bbd064ed15a89696116af59f920f9237b9113d9d846b37"} err="failed to get container status \"7f1a10aa4caf660e44bbd064ed15a89696116af59f920f9237b9113d9d846b37\": rpc error: code = NotFound desc = could not find container \"7f1a10aa4caf660e44bbd064ed15a89696116af59f920f9237b9113d9d846b37\": container with ID starting with 7f1a10aa4caf660e44bbd064ed15a89696116af59f920f9237b9113d9d846b37 not found: ID does not exist" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.128392 4998 scope.go:117] "RemoveContainer" containerID="800894309e99b901dd8864c7e28ee2fe0b7559f645f17ca3b56219578d4937e4" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.131473 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"800894309e99b901dd8864c7e28ee2fe0b7559f645f17ca3b56219578d4937e4"} err="failed to get container status \"800894309e99b901dd8864c7e28ee2fe0b7559f645f17ca3b56219578d4937e4\": rpc error: code = NotFound desc = could not find container \"800894309e99b901dd8864c7e28ee2fe0b7559f645f17ca3b56219578d4937e4\": container with ID starting with 800894309e99b901dd8864c7e28ee2fe0b7559f645f17ca3b56219578d4937e4 not found: ID does not exist" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.145000 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.191838 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.295020 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkf7v\" (UniqueName: \"kubernetes.io/projected/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-kube-api-access-fkf7v\") pod \"glance-default-internal-api-0\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.295148 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.295195 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.295233 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.295264 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.295289 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.295332 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-logs\") pod \"glance-default-internal-api-0\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.295345 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.398781 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkf7v\" (UniqueName: \"kubernetes.io/projected/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-kube-api-access-fkf7v\") pod \"glance-default-internal-api-0\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.399069 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.399105 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.399129 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.399155 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.399176 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.399210 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-logs\") pod \"glance-default-internal-api-0\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.399223 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.402080 4998 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.404827 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.405352 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-logs\") pod \"glance-default-internal-api-0\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.411110 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.411837 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.426364 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.426372 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.439687 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkf7v\" (UniqueName: \"kubernetes.io/projected/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-kube-api-access-fkf7v\") pod \"glance-default-internal-api-0\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.541021 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.721777 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.742216 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.775902 4998 generic.go:334] "Generic (PLEG): container finished" podID="855663b4-d4de-4ca0-811c-44d4def02706" containerID="2c63eb0b5ccdeff1d9234d2c6cc21f2aab83a23b017acd58cc9a301f5ca959c6" exitCode=0 Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.775967 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"855663b4-d4de-4ca0-811c-44d4def02706","Type":"ContainerDied","Data":"2c63eb0b5ccdeff1d9234d2c6cc21f2aab83a23b017acd58cc9a301f5ca959c6"} Dec 03 16:25:38 crc kubenswrapper[4998]: I1203 16:25:38.798800 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 16:25:38 crc kubenswrapper[4998]: W1203 16:25:38.812318 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d7de5e1_fc35_4b0c_8b1d_509c30d11939.slice/crio-db56ab4a8d6a9bf092191e5b57c7957937d3e81cd7afb128942a1e1ce461120a WatchSource:0}: Error finding container db56ab4a8d6a9bf092191e5b57c7957937d3e81cd7afb128942a1e1ce461120a: Status 404 returned error can't find the container with id db56ab4a8d6a9bf092191e5b57c7957937d3e81cd7afb128942a1e1ce461120a Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.113291 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.239356 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/855663b4-d4de-4ca0-811c-44d4def02706-scripts\") pod \"855663b4-d4de-4ca0-811c-44d4def02706\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.239434 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czzv5\" (UniqueName: \"kubernetes.io/projected/855663b4-d4de-4ca0-811c-44d4def02706-kube-api-access-czzv5\") pod \"855663b4-d4de-4ca0-811c-44d4def02706\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.239552 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/855663b4-d4de-4ca0-811c-44d4def02706-combined-ca-bundle\") pod \"855663b4-d4de-4ca0-811c-44d4def02706\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.239615 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/855663b4-d4de-4ca0-811c-44d4def02706-config-data\") pod \"855663b4-d4de-4ca0-811c-44d4def02706\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.239651 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/855663b4-d4de-4ca0-811c-44d4def02706-run-httpd\") pod \"855663b4-d4de-4ca0-811c-44d4def02706\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.239692 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/855663b4-d4de-4ca0-811c-44d4def02706-log-httpd\") pod \"855663b4-d4de-4ca0-811c-44d4def02706\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.239747 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/855663b4-d4de-4ca0-811c-44d4def02706-sg-core-conf-yaml\") pod \"855663b4-d4de-4ca0-811c-44d4def02706\" (UID: \"855663b4-d4de-4ca0-811c-44d4def02706\") " Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.242996 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/855663b4-d4de-4ca0-811c-44d4def02706-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "855663b4-d4de-4ca0-811c-44d4def02706" (UID: "855663b4-d4de-4ca0-811c-44d4def02706"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.243168 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/855663b4-d4de-4ca0-811c-44d4def02706-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "855663b4-d4de-4ca0-811c-44d4def02706" (UID: "855663b4-d4de-4ca0-811c-44d4def02706"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.249468 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/855663b4-d4de-4ca0-811c-44d4def02706-kube-api-access-czzv5" (OuterVolumeSpecName: "kube-api-access-czzv5") pod "855663b4-d4de-4ca0-811c-44d4def02706" (UID: "855663b4-d4de-4ca0-811c-44d4def02706"). InnerVolumeSpecName "kube-api-access-czzv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.250932 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/855663b4-d4de-4ca0-811c-44d4def02706-scripts" (OuterVolumeSpecName: "scripts") pod "855663b4-d4de-4ca0-811c-44d4def02706" (UID: "855663b4-d4de-4ca0-811c-44d4def02706"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.263976 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.305708 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/855663b4-d4de-4ca0-811c-44d4def02706-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "855663b4-d4de-4ca0-811c-44d4def02706" (UID: "855663b4-d4de-4ca0-811c-44d4def02706"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.343919 4998 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/855663b4-d4de-4ca0-811c-44d4def02706-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.344206 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/855663b4-d4de-4ca0-811c-44d4def02706-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.344315 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czzv5\" (UniqueName: \"kubernetes.io/projected/855663b4-d4de-4ca0-811c-44d4def02706-kube-api-access-czzv5\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.344414 4998 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/855663b4-d4de-4ca0-811c-44d4def02706-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.344468 4998 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/855663b4-d4de-4ca0-811c-44d4def02706-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.506877 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/855663b4-d4de-4ca0-811c-44d4def02706-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "855663b4-d4de-4ca0-811c-44d4def02706" (UID: "855663b4-d4de-4ca0-811c-44d4def02706"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.524938 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.524981 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.525643 4998 scope.go:117] "RemoveContainer" containerID="b657ed22e9096f43b7242cbed330ab7de4b3e3c8d87141fb5d8555f0560c2457" Dec 03 16:25:39 crc kubenswrapper[4998]: E1203 16:25:39.526005 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(a009aeee-c849-4c00-b41b-3cf5e36a75bb)\"" pod="openstack/watcher-decision-engine-0" podUID="a009aeee-c849-4c00-b41b-3cf5e36a75bb" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.556607 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/855663b4-d4de-4ca0-811c-44d4def02706-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.582004 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/855663b4-d4de-4ca0-811c-44d4def02706-config-data" (OuterVolumeSpecName: "config-data") pod "855663b4-d4de-4ca0-811c-44d4def02706" (UID: "855663b4-d4de-4ca0-811c-44d4def02706"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.628046 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.658866 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/855663b4-d4de-4ca0-811c-44d4def02706-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.696491 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ee49423-2774-4d51-ae4f-8cea1775efef" path="/var/lib/kubelet/pods/1ee49423-2774-4d51-ae4f-8cea1775efef/volumes" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.697293 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f32474ed-1700-42c9-b87d-ffc1f8445ea6" path="/var/lib/kubelet/pods/f32474ed-1700-42c9-b87d-ffc1f8445ea6/volumes" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.835692 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"855663b4-d4de-4ca0-811c-44d4def02706","Type":"ContainerDied","Data":"3b776860dd5eb21a0151994321cc0d35a66bb4babdf3376145b3fd6ebef47320"} Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.836100 4998 scope.go:117] "RemoveContainer" containerID="f75e25cdac20077210bd49dc9952da9b52dc8acf1d5253701e494daa9d79e244" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.836138 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.842507 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1d7de5e1-fc35-4b0c-8b1d-509c30d11939","Type":"ContainerStarted","Data":"db56ab4a8d6a9bf092191e5b57c7957937d3e81cd7afb128942a1e1ce461120a"} Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.845497 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4","Type":"ContainerStarted","Data":"585b6021eb6b0fe8eefab2d6ddbb6af3eba05c312be8c6d32d8926b631ed91ca"} Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.868797 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.880373 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.902135 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:25:39 crc kubenswrapper[4998]: E1203 16:25:39.902610 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="855663b4-d4de-4ca0-811c-44d4def02706" containerName="sg-core" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.902622 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="855663b4-d4de-4ca0-811c-44d4def02706" containerName="sg-core" Dec 03 16:25:39 crc kubenswrapper[4998]: E1203 16:25:39.902646 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="855663b4-d4de-4ca0-811c-44d4def02706" containerName="ceilometer-central-agent" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.902652 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="855663b4-d4de-4ca0-811c-44d4def02706" containerName="ceilometer-central-agent" Dec 03 16:25:39 crc kubenswrapper[4998]: E1203 16:25:39.902663 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="855663b4-d4de-4ca0-811c-44d4def02706" containerName="proxy-httpd" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.902670 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="855663b4-d4de-4ca0-811c-44d4def02706" containerName="proxy-httpd" Dec 03 16:25:39 crc kubenswrapper[4998]: E1203 16:25:39.902682 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="855663b4-d4de-4ca0-811c-44d4def02706" containerName="ceilometer-notification-agent" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.902688 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="855663b4-d4de-4ca0-811c-44d4def02706" containerName="ceilometer-notification-agent" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.902880 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="855663b4-d4de-4ca0-811c-44d4def02706" containerName="ceilometer-notification-agent" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.902891 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="855663b4-d4de-4ca0-811c-44d4def02706" containerName="proxy-httpd" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.902904 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="855663b4-d4de-4ca0-811c-44d4def02706" containerName="ceilometer-central-agent" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.902916 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="855663b4-d4de-4ca0-811c-44d4def02706" containerName="sg-core" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.904484 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.908727 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.919123 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.925228 4998 scope.go:117] "RemoveContainer" containerID="d467c7642ab7b67ec0bb6e94bf7954b5748a94ef142fb733a8fe87b353b6838d" Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.925669 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:25:39 crc kubenswrapper[4998]: I1203 16:25:39.966719 4998 scope.go:117] "RemoveContainer" containerID="2c63eb0b5ccdeff1d9234d2c6cc21f2aab83a23b017acd58cc9a301f5ca959c6" Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.000365 4998 scope.go:117] "RemoveContainer" containerID="bd642294c5a767432620ab6245505b9509dcd42d5a1e0c7143d68ee5baa5b382" Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.068126 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " pod="openstack/ceilometer-0" Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.068194 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kvps\" (UniqueName: \"kubernetes.io/projected/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-kube-api-access-7kvps\") pod \"ceilometer-0\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " pod="openstack/ceilometer-0" Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.068241 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-config-data\") pod \"ceilometer-0\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " pod="openstack/ceilometer-0" Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.068411 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-scripts\") pod \"ceilometer-0\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " pod="openstack/ceilometer-0" Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.068483 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-run-httpd\") pod \"ceilometer-0\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " pod="openstack/ceilometer-0" Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.068513 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " pod="openstack/ceilometer-0" Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.068772 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-log-httpd\") pod \"ceilometer-0\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " pod="openstack/ceilometer-0" Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.170741 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " pod="openstack/ceilometer-0" Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.170877 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kvps\" (UniqueName: \"kubernetes.io/projected/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-kube-api-access-7kvps\") pod \"ceilometer-0\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " pod="openstack/ceilometer-0" Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.170898 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-config-data\") pod \"ceilometer-0\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " pod="openstack/ceilometer-0" Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.170938 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-scripts\") pod \"ceilometer-0\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " pod="openstack/ceilometer-0" Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.170963 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-run-httpd\") pod \"ceilometer-0\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " pod="openstack/ceilometer-0" Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.170979 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " pod="openstack/ceilometer-0" Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.171034 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-log-httpd\") pod \"ceilometer-0\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " pod="openstack/ceilometer-0" Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.171424 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-log-httpd\") pod \"ceilometer-0\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " pod="openstack/ceilometer-0" Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.172543 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-run-httpd\") pod \"ceilometer-0\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " pod="openstack/ceilometer-0" Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.186300 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " pod="openstack/ceilometer-0" Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.204618 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-scripts\") pod \"ceilometer-0\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " pod="openstack/ceilometer-0" Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.204634 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " pod="openstack/ceilometer-0" Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.204834 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-config-data\") pod \"ceilometer-0\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " pod="openstack/ceilometer-0" Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.212366 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kvps\" (UniqueName: \"kubernetes.io/projected/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-kube-api-access-7kvps\") pod \"ceilometer-0\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " pod="openstack/ceilometer-0" Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.232109 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.757963 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.827087 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcb95679f-wbxs5"] Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.827680 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" podUID="c61c77db-8343-4090-ad97-06d7814f475c" containerName="dnsmasq-dns" containerID="cri-o://76284935ee15058c4a4f41334f6b19b2e9ff45c83838ecff73ad8f612b579462" gracePeriod=10 Dec 03 16:25:40 crc kubenswrapper[4998]: I1203 16:25:40.871686 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1d7de5e1-fc35-4b0c-8b1d-509c30d11939","Type":"ContainerStarted","Data":"89c33d9ec7f919bebf6bcd6a1e9679473e0531814045827c4bca1cb28a037a00"} Dec 03 16:25:41 crc kubenswrapper[4998]: I1203 16:25:41.262816 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:25:41 crc kubenswrapper[4998]: I1203 16:25:41.785502 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="855663b4-d4de-4ca0-811c-44d4def02706" path="/var/lib/kubelet/pods/855663b4-d4de-4ca0-811c-44d4def02706/volumes" Dec 03 16:25:41 crc kubenswrapper[4998]: I1203 16:25:41.900149 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6ae06e4-ee1a-4519-beb5-23e71725f1a8","Type":"ContainerStarted","Data":"1bdf81685a0b50e81a3acf0d4b3fc76930ae0be956c5606075a3c12c42ca8533"} Dec 03 16:25:41 crc kubenswrapper[4998]: I1203 16:25:41.900503 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6ae06e4-ee1a-4519-beb5-23e71725f1a8","Type":"ContainerStarted","Data":"fb3cdf5f47ef30e84e7c31ef8f9e187833b5e42c5f6260b01de9b49fceb214da"} Dec 03 16:25:41 crc kubenswrapper[4998]: I1203 16:25:41.908490 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1d7de5e1-fc35-4b0c-8b1d-509c30d11939","Type":"ContainerStarted","Data":"4e719b9f4645604bbdc077703eb3389a2f670a438a0887e292ab1a908019f135"} Dec 03 16:25:41 crc kubenswrapper[4998]: I1203 16:25:41.916153 4998 generic.go:334] "Generic (PLEG): container finished" podID="c61c77db-8343-4090-ad97-06d7814f475c" containerID="76284935ee15058c4a4f41334f6b19b2e9ff45c83838ecff73ad8f612b579462" exitCode=0 Dec 03 16:25:41 crc kubenswrapper[4998]: I1203 16:25:41.916231 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" event={"ID":"c61c77db-8343-4090-ad97-06d7814f475c","Type":"ContainerDied","Data":"76284935ee15058c4a4f41334f6b19b2e9ff45c83838ecff73ad8f612b579462"} Dec 03 16:25:41 crc kubenswrapper[4998]: I1203 16:25:41.925678 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4","Type":"ContainerStarted","Data":"1657b6f33a5e6e5d2f68ab2fa8bc506670d5df5ca49603145704a3b5d0967132"} Dec 03 16:25:41 crc kubenswrapper[4998]: I1203 16:25:41.940248 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.9402261 podStartE2EDuration="4.9402261s" podCreationTimestamp="2025-12-03 16:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:25:41.937184386 +0000 UTC m=+1320.548884619" watchObservedRunningTime="2025-12-03 16:25:41.9402261 +0000 UTC m=+1320.551926323" Dec 03 16:25:42 crc kubenswrapper[4998]: I1203 16:25:42.948118 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4","Type":"ContainerStarted","Data":"9b5be307c8ff3af1a80772f8810dddbc54affcdf58a7fb498209aac41357a527"} Dec 03 16:25:42 crc kubenswrapper[4998]: I1203 16:25:42.973164 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.973144226 podStartE2EDuration="4.973144226s" podCreationTimestamp="2025-12-03 16:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:25:42.964166648 +0000 UTC m=+1321.575866871" watchObservedRunningTime="2025-12-03 16:25:42.973144226 +0000 UTC m=+1321.584844449" Dec 03 16:25:44 crc kubenswrapper[4998]: I1203 16:25:44.003820 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:44 crc kubenswrapper[4998]: I1203 16:25:44.008691 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-59d5c55c95-qdh4t" Dec 03 16:25:44 crc kubenswrapper[4998]: I1203 16:25:44.405199 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 16:25:44 crc kubenswrapper[4998]: I1203 16:25:44.589249 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:25:48 crc kubenswrapper[4998]: I1203 16:25:48.192045 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 16:25:48 crc kubenswrapper[4998]: I1203 16:25:48.192893 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 16:25:48 crc kubenswrapper[4998]: I1203 16:25:48.234709 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 16:25:48 crc kubenswrapper[4998]: I1203 16:25:48.256321 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 16:25:48 crc kubenswrapper[4998]: I1203 16:25:48.742884 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 16:25:48 crc kubenswrapper[4998]: I1203 16:25:48.743165 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 16:25:48 crc kubenswrapper[4998]: I1203 16:25:48.778802 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 16:25:48 crc kubenswrapper[4998]: I1203 16:25:48.784336 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 16:25:48 crc kubenswrapper[4998]: I1203 16:25:48.967065 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.067888 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" event={"ID":"c61c77db-8343-4090-ad97-06d7814f475c","Type":"ContainerDied","Data":"50d1c77b3705953dd177fbf55913122d319a25fa662e9b970f3c4201d941e15e"} Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.068321 4998 scope.go:117] "RemoveContainer" containerID="76284935ee15058c4a4f41334f6b19b2e9ff45c83838ecff73ad8f612b579462" Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.068834 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.068881 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.069456 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.071204 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.071244 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.081799 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-ovsdbserver-nb\") pod \"c61c77db-8343-4090-ad97-06d7814f475c\" (UID: \"c61c77db-8343-4090-ad97-06d7814f475c\") " Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.081851 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-ovsdbserver-sb\") pod \"c61c77db-8343-4090-ad97-06d7814f475c\" (UID: \"c61c77db-8343-4090-ad97-06d7814f475c\") " Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.081970 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-config\") pod \"c61c77db-8343-4090-ad97-06d7814f475c\" (UID: \"c61c77db-8343-4090-ad97-06d7814f475c\") " Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.082022 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-dns-svc\") pod \"c61c77db-8343-4090-ad97-06d7814f475c\" (UID: \"c61c77db-8343-4090-ad97-06d7814f475c\") " Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.082076 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9c7t6\" (UniqueName: \"kubernetes.io/projected/c61c77db-8343-4090-ad97-06d7814f475c-kube-api-access-9c7t6\") pod \"c61c77db-8343-4090-ad97-06d7814f475c\" (UID: \"c61c77db-8343-4090-ad97-06d7814f475c\") " Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.082109 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-dns-swift-storage-0\") pod \"c61c77db-8343-4090-ad97-06d7814f475c\" (UID: \"c61c77db-8343-4090-ad97-06d7814f475c\") " Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.107767 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c61c77db-8343-4090-ad97-06d7814f475c-kube-api-access-9c7t6" (OuterVolumeSpecName: "kube-api-access-9c7t6") pod "c61c77db-8343-4090-ad97-06d7814f475c" (UID: "c61c77db-8343-4090-ad97-06d7814f475c"). InnerVolumeSpecName "kube-api-access-9c7t6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.164570 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c61c77db-8343-4090-ad97-06d7814f475c" (UID: "c61c77db-8343-4090-ad97-06d7814f475c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.174769 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c61c77db-8343-4090-ad97-06d7814f475c" (UID: "c61c77db-8343-4090-ad97-06d7814f475c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.181936 4998 scope.go:117] "RemoveContainer" containerID="bbb23e2206a12773508000e32cddaf2d381f59568d74f7a45ac7ddec5b40bf71" Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.183439 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c61c77db-8343-4090-ad97-06d7814f475c" (UID: "c61c77db-8343-4090-ad97-06d7814f475c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.183597 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-dns-swift-storage-0\") pod \"c61c77db-8343-4090-ad97-06d7814f475c\" (UID: \"c61c77db-8343-4090-ad97-06d7814f475c\") " Dec 03 16:25:49 crc kubenswrapper[4998]: W1203 16:25:49.183666 4998 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/c61c77db-8343-4090-ad97-06d7814f475c/volumes/kubernetes.io~configmap/dns-swift-storage-0 Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.183680 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c61c77db-8343-4090-ad97-06d7814f475c" (UID: "c61c77db-8343-4090-ad97-06d7814f475c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.185660 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9c7t6\" (UniqueName: \"kubernetes.io/projected/c61c77db-8343-4090-ad97-06d7814f475c-kube-api-access-9c7t6\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.185677 4998 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.185686 4998 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.185696 4998 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.212819 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c61c77db-8343-4090-ad97-06d7814f475c" (UID: "c61c77db-8343-4090-ad97-06d7814f475c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.237209 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-config" (OuterVolumeSpecName: "config") pod "c61c77db-8343-4090-ad97-06d7814f475c" (UID: "c61c77db-8343-4090-ad97-06d7814f475c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.240593 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-fcb95679f-wbxs5" podUID="c61c77db-8343-4090-ad97-06d7814f475c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.169:5353: i/o timeout" Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.287150 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.287320 4998 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c61c77db-8343-4090-ad97-06d7814f475c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.402571 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcb95679f-wbxs5"] Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.410037 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-fcb95679f-wbxs5"] Dec 03 16:25:49 crc kubenswrapper[4998]: I1203 16:25:49.688012 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c61c77db-8343-4090-ad97-06d7814f475c" path="/var/lib/kubelet/pods/c61c77db-8343-4090-ad97-06d7814f475c/volumes" Dec 03 16:25:50 crc kubenswrapper[4998]: I1203 16:25:50.078885 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6ae06e4-ee1a-4519-beb5-23e71725f1a8","Type":"ContainerStarted","Data":"d6a6766732ac2372806808fa1bc057c7492cc31b9a26c7cdde36bd7df397d3ec"} Dec 03 16:25:50 crc kubenswrapper[4998]: I1203 16:25:50.083319 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"041471bf-1760-4fb8-ac9a-81dc1dd7defa","Type":"ContainerStarted","Data":"0fcf0b091acac2d34a16debaa0b5069ed2b1199769bf912ca13eba6672f96aaf"} Dec 03 16:25:50 crc kubenswrapper[4998]: I1203 16:25:50.099913 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.867992577 podStartE2EDuration="25.099894197s" podCreationTimestamp="2025-12-03 16:25:25 +0000 UTC" firstStartedPulling="2025-12-03 16:25:26.73635062 +0000 UTC m=+1305.348050833" lastFinishedPulling="2025-12-03 16:25:48.96825223 +0000 UTC m=+1327.579952453" observedRunningTime="2025-12-03 16:25:50.096463984 +0000 UTC m=+1328.708164207" watchObservedRunningTime="2025-12-03 16:25:50.099894197 +0000 UTC m=+1328.711594420" Dec 03 16:25:50 crc kubenswrapper[4998]: I1203 16:25:50.678586 4998 scope.go:117] "RemoveContainer" containerID="b657ed22e9096f43b7242cbed330ab7de4b3e3c8d87141fb5d8555f0560c2457" Dec 03 16:25:50 crc kubenswrapper[4998]: E1203 16:25:50.679130 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-decision-engine\" with CrashLoopBackOff: \"back-off 20s restarting failed container=watcher-decision-engine pod=watcher-decision-engine-0_openstack(a009aeee-c849-4c00-b41b-3cf5e36a75bb)\"" pod="openstack/watcher-decision-engine-0" podUID="a009aeee-c849-4c00-b41b-3cf5e36a75bb" Dec 03 16:25:51 crc kubenswrapper[4998]: I1203 16:25:51.093146 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6ae06e4-ee1a-4519-beb5-23e71725f1a8","Type":"ContainerStarted","Data":"6c1bbc95fec653f64e043760f67069f83aba559b1b70015ca08fd74c0e1fa2ab"} Dec 03 16:25:51 crc kubenswrapper[4998]: I1203 16:25:51.093196 4998 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 16:25:51 crc kubenswrapper[4998]: I1203 16:25:51.093878 4998 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 16:25:51 crc kubenswrapper[4998]: I1203 16:25:51.093221 4998 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 16:25:51 crc kubenswrapper[4998]: I1203 16:25:51.094066 4998 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 16:25:53 crc kubenswrapper[4998]: I1203 16:25:53.927426 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 16:25:53 crc kubenswrapper[4998]: I1203 16:25:53.928047 4998 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 16:25:53 crc kubenswrapper[4998]: I1203 16:25:53.930467 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 16:25:53 crc kubenswrapper[4998]: I1203 16:25:53.998153 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 16:25:53 crc kubenswrapper[4998]: I1203 16:25:53.998265 4998 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 16:25:54 crc kubenswrapper[4998]: I1203 16:25:54.000165 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 16:25:54 crc kubenswrapper[4998]: I1203 16:25:54.132183 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d6ae06e4-ee1a-4519-beb5-23e71725f1a8" containerName="ceilometer-central-agent" containerID="cri-o://1bdf81685a0b50e81a3acf0d4b3fc76930ae0be956c5606075a3c12c42ca8533" gracePeriod=30 Dec 03 16:25:54 crc kubenswrapper[4998]: I1203 16:25:54.132473 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6ae06e4-ee1a-4519-beb5-23e71725f1a8","Type":"ContainerStarted","Data":"73394b01935e50c7874fd33ad15b0c717e15f262af32534878e62bb641f693ed"} Dec 03 16:25:54 crc kubenswrapper[4998]: I1203 16:25:54.132946 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 16:25:54 crc kubenswrapper[4998]: I1203 16:25:54.133167 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d6ae06e4-ee1a-4519-beb5-23e71725f1a8" containerName="proxy-httpd" containerID="cri-o://73394b01935e50c7874fd33ad15b0c717e15f262af32534878e62bb641f693ed" gracePeriod=30 Dec 03 16:25:54 crc kubenswrapper[4998]: I1203 16:25:54.133216 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d6ae06e4-ee1a-4519-beb5-23e71725f1a8" containerName="sg-core" containerID="cri-o://6c1bbc95fec653f64e043760f67069f83aba559b1b70015ca08fd74c0e1fa2ab" gracePeriod=30 Dec 03 16:25:54 crc kubenswrapper[4998]: I1203 16:25:54.133259 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d6ae06e4-ee1a-4519-beb5-23e71725f1a8" containerName="ceilometer-notification-agent" containerID="cri-o://d6a6766732ac2372806808fa1bc057c7492cc31b9a26c7cdde36bd7df397d3ec" gracePeriod=30 Dec 03 16:25:54 crc kubenswrapper[4998]: I1203 16:25:54.164671 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.790682973 podStartE2EDuration="15.164653284s" podCreationTimestamp="2025-12-03 16:25:39 +0000 UTC" firstStartedPulling="2025-12-03 16:25:41.279454707 +0000 UTC m=+1319.891154940" lastFinishedPulling="2025-12-03 16:25:53.653425018 +0000 UTC m=+1332.265125251" observedRunningTime="2025-12-03 16:25:54.153818081 +0000 UTC m=+1332.765518314" watchObservedRunningTime="2025-12-03 16:25:54.164653284 +0000 UTC m=+1332.776353517" Dec 03 16:25:54 crc kubenswrapper[4998]: E1203 16:25:54.478993 4998 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6ae06e4_ee1a_4519_beb5_23e71725f1a8.slice/crio-73394b01935e50c7874fd33ad15b0c717e15f262af32534878e62bb641f693ed.scope\": RecentStats: unable to find data in memory cache]" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.163609 4998 generic.go:334] "Generic (PLEG): container finished" podID="a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf" containerID="a02f38a8cb552441e6687ec55547b52c4b41a36f038a3ee855fa5659d41077a3" exitCode=137 Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.163816 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf","Type":"ContainerDied","Data":"a02f38a8cb552441e6687ec55547b52c4b41a36f038a3ee855fa5659d41077a3"} Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.171052 4998 generic.go:334] "Generic (PLEG): container finished" podID="d6ae06e4-ee1a-4519-beb5-23e71725f1a8" containerID="73394b01935e50c7874fd33ad15b0c717e15f262af32534878e62bb641f693ed" exitCode=0 Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.171089 4998 generic.go:334] "Generic (PLEG): container finished" podID="d6ae06e4-ee1a-4519-beb5-23e71725f1a8" containerID="6c1bbc95fec653f64e043760f67069f83aba559b1b70015ca08fd74c0e1fa2ab" exitCode=2 Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.171102 4998 generic.go:334] "Generic (PLEG): container finished" podID="d6ae06e4-ee1a-4519-beb5-23e71725f1a8" containerID="d6a6766732ac2372806808fa1bc057c7492cc31b9a26c7cdde36bd7df397d3ec" exitCode=0 Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.171112 4998 generic.go:334] "Generic (PLEG): container finished" podID="d6ae06e4-ee1a-4519-beb5-23e71725f1a8" containerID="1bdf81685a0b50e81a3acf0d4b3fc76930ae0be956c5606075a3c12c42ca8533" exitCode=0 Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.171134 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6ae06e4-ee1a-4519-beb5-23e71725f1a8","Type":"ContainerDied","Data":"73394b01935e50c7874fd33ad15b0c717e15f262af32534878e62bb641f693ed"} Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.171164 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6ae06e4-ee1a-4519-beb5-23e71725f1a8","Type":"ContainerDied","Data":"6c1bbc95fec653f64e043760f67069f83aba559b1b70015ca08fd74c0e1fa2ab"} Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.171177 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6ae06e4-ee1a-4519-beb5-23e71725f1a8","Type":"ContainerDied","Data":"d6a6766732ac2372806808fa1bc057c7492cc31b9a26c7cdde36bd7df397d3ec"} Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.171189 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6ae06e4-ee1a-4519-beb5-23e71725f1a8","Type":"ContainerDied","Data":"1bdf81685a0b50e81a3acf0d4b3fc76930ae0be956c5606075a3c12c42ca8533"} Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.222607 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.353146 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-config-data\") pod \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.353222 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-log-httpd\") pod \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.353250 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kvps\" (UniqueName: \"kubernetes.io/projected/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-kube-api-access-7kvps\") pod \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.353300 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-sg-core-conf-yaml\") pod \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.353338 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-run-httpd\") pod \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.353379 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-combined-ca-bundle\") pod \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.353404 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-scripts\") pod \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\" (UID: \"d6ae06e4-ee1a-4519-beb5-23e71725f1a8\") " Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.355814 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d6ae06e4-ee1a-4519-beb5-23e71725f1a8" (UID: "d6ae06e4-ee1a-4519-beb5-23e71725f1a8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.356166 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d6ae06e4-ee1a-4519-beb5-23e71725f1a8" (UID: "d6ae06e4-ee1a-4519-beb5-23e71725f1a8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.360299 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-kube-api-access-7kvps" (OuterVolumeSpecName: "kube-api-access-7kvps") pod "d6ae06e4-ee1a-4519-beb5-23e71725f1a8" (UID: "d6ae06e4-ee1a-4519-beb5-23e71725f1a8"). InnerVolumeSpecName "kube-api-access-7kvps". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.361723 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-scripts" (OuterVolumeSpecName: "scripts") pod "d6ae06e4-ee1a-4519-beb5-23e71725f1a8" (UID: "d6ae06e4-ee1a-4519-beb5-23e71725f1a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.376907 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.429568 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d6ae06e4-ee1a-4519-beb5-23e71725f1a8" (UID: "d6ae06e4-ee1a-4519-beb5-23e71725f1a8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.441700 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6ae06e4-ee1a-4519-beb5-23e71725f1a8" (UID: "d6ae06e4-ee1a-4519-beb5-23e71725f1a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.455274 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-scripts\") pod \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.455376 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-config-data\") pod \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.455412 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-logs\") pod \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.455438 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-config-data-custom\") pod \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.455465 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-etc-machine-id\") pod \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.455490 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7gkh\" (UniqueName: \"kubernetes.io/projected/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-kube-api-access-v7gkh\") pod \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.455509 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-combined-ca-bundle\") pod \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\" (UID: \"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf\") " Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.456169 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.456189 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.456201 4998 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.456213 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kvps\" (UniqueName: \"kubernetes.io/projected/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-kube-api-access-7kvps\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.456228 4998 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.456237 4998 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.457649 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-logs" (OuterVolumeSpecName: "logs") pod "a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf" (UID: "a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.458996 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf" (UID: "a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.462101 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-scripts" (OuterVolumeSpecName: "scripts") pod "a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf" (UID: "a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.465728 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf" (UID: "a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.470371 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-kube-api-access-v7gkh" (OuterVolumeSpecName: "kube-api-access-v7gkh") pod "a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf" (UID: "a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf"). InnerVolumeSpecName "kube-api-access-v7gkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.488468 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf" (UID: "a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.500737 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-config-data" (OuterVolumeSpecName: "config-data") pod "d6ae06e4-ee1a-4519-beb5-23e71725f1a8" (UID: "d6ae06e4-ee1a-4519-beb5-23e71725f1a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.538610 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-config-data" (OuterVolumeSpecName: "config-data") pod "a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf" (UID: "a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.559373 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.559411 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.559427 4998 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-logs\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.559437 4998 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.559448 4998 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.559457 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7gkh\" (UniqueName: \"kubernetes.io/projected/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-kube-api-access-v7gkh\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.559467 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:55 crc kubenswrapper[4998]: I1203 16:25:55.559476 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6ae06e4-ee1a-4519-beb5-23e71725f1a8-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.183728 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf","Type":"ContainerDied","Data":"e4e63573804de55c264f465658b61f086c9b0740e45b8c509f6f0e81a143e022"} Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.183824 4998 scope.go:117] "RemoveContainer" containerID="a02f38a8cb552441e6687ec55547b52c4b41a36f038a3ee855fa5659d41077a3" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.184922 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.187548 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6ae06e4-ee1a-4519-beb5-23e71725f1a8","Type":"ContainerDied","Data":"fb3cdf5f47ef30e84e7c31ef8f9e187833b5e42c5f6260b01de9b49fceb214da"} Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.187672 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.218041 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.225696 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.233359 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.249550 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.252449 4998 scope.go:117] "RemoveContainer" containerID="4cd5621bd9999f8912e9f8dac7be546d02d473e54439ae066cb6d6d394507220" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.259423 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 16:25:56 crc kubenswrapper[4998]: E1203 16:25:56.259972 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c61c77db-8343-4090-ad97-06d7814f475c" containerName="dnsmasq-dns" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.259992 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="c61c77db-8343-4090-ad97-06d7814f475c" containerName="dnsmasq-dns" Dec 03 16:25:56 crc kubenswrapper[4998]: E1203 16:25:56.260012 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c61c77db-8343-4090-ad97-06d7814f475c" containerName="init" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.260021 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="c61c77db-8343-4090-ad97-06d7814f475c" containerName="init" Dec 03 16:25:56 crc kubenswrapper[4998]: E1203 16:25:56.260029 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6ae06e4-ee1a-4519-beb5-23e71725f1a8" containerName="proxy-httpd" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.260037 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6ae06e4-ee1a-4519-beb5-23e71725f1a8" containerName="proxy-httpd" Dec 03 16:25:56 crc kubenswrapper[4998]: E1203 16:25:56.260053 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6ae06e4-ee1a-4519-beb5-23e71725f1a8" containerName="ceilometer-notification-agent" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.260060 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6ae06e4-ee1a-4519-beb5-23e71725f1a8" containerName="ceilometer-notification-agent" Dec 03 16:25:56 crc kubenswrapper[4998]: E1203 16:25:56.260080 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6ae06e4-ee1a-4519-beb5-23e71725f1a8" containerName="sg-core" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.260088 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6ae06e4-ee1a-4519-beb5-23e71725f1a8" containerName="sg-core" Dec 03 16:25:56 crc kubenswrapper[4998]: E1203 16:25:56.260111 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6ae06e4-ee1a-4519-beb5-23e71725f1a8" containerName="ceilometer-central-agent" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.260119 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6ae06e4-ee1a-4519-beb5-23e71725f1a8" containerName="ceilometer-central-agent" Dec 03 16:25:56 crc kubenswrapper[4998]: E1203 16:25:56.260129 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf" containerName="cinder-api" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.260136 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf" containerName="cinder-api" Dec 03 16:25:56 crc kubenswrapper[4998]: E1203 16:25:56.260153 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf" containerName="cinder-api-log" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.260160 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf" containerName="cinder-api-log" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.260361 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6ae06e4-ee1a-4519-beb5-23e71725f1a8" containerName="ceilometer-central-agent" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.260378 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6ae06e4-ee1a-4519-beb5-23e71725f1a8" containerName="ceilometer-notification-agent" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.260388 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="c61c77db-8343-4090-ad97-06d7814f475c" containerName="dnsmasq-dns" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.260397 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf" containerName="cinder-api" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.260412 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf" containerName="cinder-api-log" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.260435 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6ae06e4-ee1a-4519-beb5-23e71725f1a8" containerName="sg-core" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.260453 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6ae06e4-ee1a-4519-beb5-23e71725f1a8" containerName="proxy-httpd" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.262549 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.269201 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.269249 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.269288 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.269730 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.276448 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.278944 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.283453 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.285973 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.286197 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.293924 4998 scope.go:117] "RemoveContainer" containerID="73394b01935e50c7874fd33ad15b0c717e15f262af32534878e62bb641f693ed" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.314583 4998 scope.go:117] "RemoveContainer" containerID="6c1bbc95fec653f64e043760f67069f83aba559b1b70015ca08fd74c0e1fa2ab" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.335246 4998 scope.go:117] "RemoveContainer" containerID="d6a6766732ac2372806808fa1bc057c7492cc31b9a26c7cdde36bd7df397d3ec" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.364337 4998 scope.go:117] "RemoveContainer" containerID="1bdf81685a0b50e81a3acf0d4b3fc76930ae0be956c5606075a3c12c42ca8533" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.372795 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzqlg\" (UniqueName: \"kubernetes.io/projected/937de507-f2cb-4b8c-8b86-cfc929f23248-kube-api-access-tzqlg\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.372920 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/937de507-f2cb-4b8c-8b86-cfc929f23248-config-data-custom\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.372990 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/937de507-f2cb-4b8c-8b86-cfc929f23248-config-data\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.373599 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/937de507-f2cb-4b8c-8b86-cfc929f23248-logs\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.373654 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/937de507-f2cb-4b8c-8b86-cfc929f23248-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.373687 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/937de507-f2cb-4b8c-8b86-cfc929f23248-etc-machine-id\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.373703 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/937de507-f2cb-4b8c-8b86-cfc929f23248-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.373720 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/937de507-f2cb-4b8c-8b86-cfc929f23248-scripts\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.374186 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/937de507-f2cb-4b8c-8b86-cfc929f23248-public-tls-certs\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.475531 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " pod="openstack/ceilometer-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.475855 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/937de507-f2cb-4b8c-8b86-cfc929f23248-config-data-custom\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.475977 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/937de507-f2cb-4b8c-8b86-cfc929f23248-config-data\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.476058 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/937de507-f2cb-4b8c-8b86-cfc929f23248-logs\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.476155 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/937de507-f2cb-4b8c-8b86-cfc929f23248-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.476235 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-config-data\") pod \"ceilometer-0\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " pod="openstack/ceilometer-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.476323 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/937de507-f2cb-4b8c-8b86-cfc929f23248-etc-machine-id\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.476399 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/937de507-f2cb-4b8c-8b86-cfc929f23248-etc-machine-id\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.476410 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/937de507-f2cb-4b8c-8b86-cfc929f23248-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.476484 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/937de507-f2cb-4b8c-8b86-cfc929f23248-scripts\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.476546 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/937de507-f2cb-4b8c-8b86-cfc929f23248-logs\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.476555 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-scripts\") pod \"ceilometer-0\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " pod="openstack/ceilometer-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.476701 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-log-httpd\") pod \"ceilometer-0\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " pod="openstack/ceilometer-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.476733 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " pod="openstack/ceilometer-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.476858 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/937de507-f2cb-4b8c-8b86-cfc929f23248-public-tls-certs\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.477001 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzqlg\" (UniqueName: \"kubernetes.io/projected/937de507-f2cb-4b8c-8b86-cfc929f23248-kube-api-access-tzqlg\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.477032 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-run-httpd\") pod \"ceilometer-0\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " pod="openstack/ceilometer-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.477128 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrc4z\" (UniqueName: \"kubernetes.io/projected/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-kube-api-access-jrc4z\") pod \"ceilometer-0\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " pod="openstack/ceilometer-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.482268 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/937de507-f2cb-4b8c-8b86-cfc929f23248-config-data\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.484256 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/937de507-f2cb-4b8c-8b86-cfc929f23248-public-tls-certs\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.484315 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/937de507-f2cb-4b8c-8b86-cfc929f23248-scripts\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.484420 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/937de507-f2cb-4b8c-8b86-cfc929f23248-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.484767 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/937de507-f2cb-4b8c-8b86-cfc929f23248-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.484910 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/937de507-f2cb-4b8c-8b86-cfc929f23248-config-data-custom\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.495135 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzqlg\" (UniqueName: \"kubernetes.io/projected/937de507-f2cb-4b8c-8b86-cfc929f23248-kube-api-access-tzqlg\") pod \"cinder-api-0\" (UID: \"937de507-f2cb-4b8c-8b86-cfc929f23248\") " pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.579244 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-config-data\") pod \"ceilometer-0\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " pod="openstack/ceilometer-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.579311 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-scripts\") pod \"ceilometer-0\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " pod="openstack/ceilometer-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.579331 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-log-httpd\") pod \"ceilometer-0\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " pod="openstack/ceilometer-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.579346 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " pod="openstack/ceilometer-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.579397 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-run-httpd\") pod \"ceilometer-0\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " pod="openstack/ceilometer-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.579431 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrc4z\" (UniqueName: \"kubernetes.io/projected/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-kube-api-access-jrc4z\") pod \"ceilometer-0\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " pod="openstack/ceilometer-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.579463 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " pod="openstack/ceilometer-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.580410 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-run-httpd\") pod \"ceilometer-0\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " pod="openstack/ceilometer-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.580532 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-log-httpd\") pod \"ceilometer-0\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " pod="openstack/ceilometer-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.584708 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " pod="openstack/ceilometer-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.585462 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-config-data\") pod \"ceilometer-0\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " pod="openstack/ceilometer-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.586186 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-scripts\") pod \"ceilometer-0\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " pod="openstack/ceilometer-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.593518 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " pod="openstack/ceilometer-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.596066 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrc4z\" (UniqueName: \"kubernetes.io/projected/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-kube-api-access-jrc4z\") pod \"ceilometer-0\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " pod="openstack/ceilometer-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.599521 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 16:25:56 crc kubenswrapper[4998]: I1203 16:25:56.610882 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.164838 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.216805 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-tdm2m"] Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.218863 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-tdm2m" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.224813 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb","Type":"ContainerStarted","Data":"9d17b2b0f53adc3a709aa486133f615dfde872e9ac0e2a248cb5c799030d5856"} Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.228674 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-tdm2m"] Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.283558 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.315149 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-7bf7-account-create-update-xqtj9"] Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.323527 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7bf7-account-create-update-xqtj9" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.326885 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.333699 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-8rt9f"] Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.337319 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-8rt9f" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.352860 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-7bf7-account-create-update-xqtj9"] Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.400631 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c49a0bb7-935a-4d43-bd6f-848d5e44b43f-operator-scripts\") pod \"nova-api-db-create-tdm2m\" (UID: \"c49a0bb7-935a-4d43-bd6f-848d5e44b43f\") " pod="openstack/nova-api-db-create-tdm2m" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.400681 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsh2t\" (UniqueName: \"kubernetes.io/projected/c49a0bb7-935a-4d43-bd6f-848d5e44b43f-kube-api-access-zsh2t\") pod \"nova-api-db-create-tdm2m\" (UID: \"c49a0bb7-935a-4d43-bd6f-848d5e44b43f\") " pod="openstack/nova-api-db-create-tdm2m" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.403790 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-8rt9f"] Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.430619 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-s9sg4"] Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.432176 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-s9sg4" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.447464 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-s9sg4"] Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.491268 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-4825-account-create-update-k82lm"] Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.494459 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-4825-account-create-update-k82lm" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.497051 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.503167 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5b99365-b407-4871-9930-1c7fa659c397-operator-scripts\") pod \"nova-cell0-db-create-8rt9f\" (UID: \"e5b99365-b407-4871-9930-1c7fa659c397\") " pod="openstack/nova-cell0-db-create-8rt9f" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.503513 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7a9cab4-be56-4a01-9b34-493370f1a56a-operator-scripts\") pod \"nova-api-7bf7-account-create-update-xqtj9\" (UID: \"f7a9cab4-be56-4a01-9b34-493370f1a56a\") " pod="openstack/nova-api-7bf7-account-create-update-xqtj9" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.503645 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgtnt\" (UniqueName: \"kubernetes.io/projected/f7a9cab4-be56-4a01-9b34-493370f1a56a-kube-api-access-kgtnt\") pod \"nova-api-7bf7-account-create-update-xqtj9\" (UID: \"f7a9cab4-be56-4a01-9b34-493370f1a56a\") " pod="openstack/nova-api-7bf7-account-create-update-xqtj9" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.503723 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c49a0bb7-935a-4d43-bd6f-848d5e44b43f-operator-scripts\") pod \"nova-api-db-create-tdm2m\" (UID: \"c49a0bb7-935a-4d43-bd6f-848d5e44b43f\") " pod="openstack/nova-api-db-create-tdm2m" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.503747 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsh2t\" (UniqueName: \"kubernetes.io/projected/c49a0bb7-935a-4d43-bd6f-848d5e44b43f-kube-api-access-zsh2t\") pod \"nova-api-db-create-tdm2m\" (UID: \"c49a0bb7-935a-4d43-bd6f-848d5e44b43f\") " pod="openstack/nova-api-db-create-tdm2m" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.504550 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c49a0bb7-935a-4d43-bd6f-848d5e44b43f-operator-scripts\") pod \"nova-api-db-create-tdm2m\" (UID: \"c49a0bb7-935a-4d43-bd6f-848d5e44b43f\") " pod="openstack/nova-api-db-create-tdm2m" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.514600 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-4825-account-create-update-k82lm"] Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.520724 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2v2r\" (UniqueName: \"kubernetes.io/projected/e5b99365-b407-4871-9930-1c7fa659c397-kube-api-access-f2v2r\") pod \"nova-cell0-db-create-8rt9f\" (UID: \"e5b99365-b407-4871-9930-1c7fa659c397\") " pod="openstack/nova-cell0-db-create-8rt9f" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.521030 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsh2t\" (UniqueName: \"kubernetes.io/projected/c49a0bb7-935a-4d43-bd6f-848d5e44b43f-kube-api-access-zsh2t\") pod \"nova-api-db-create-tdm2m\" (UID: \"c49a0bb7-935a-4d43-bd6f-848d5e44b43f\") " pod="openstack/nova-api-db-create-tdm2m" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.550734 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-tdm2m" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.628039 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqvrb\" (UniqueName: \"kubernetes.io/projected/0e609dd6-5dee-4791-91e7-21e3191db678-kube-api-access-cqvrb\") pod \"nova-cell0-4825-account-create-update-k82lm\" (UID: \"0e609dd6-5dee-4791-91e7-21e3191db678\") " pod="openstack/nova-cell0-4825-account-create-update-k82lm" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.628084 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99031ef1-ff62-4058-bf84-5c2240fbce40-operator-scripts\") pod \"nova-cell1-db-create-s9sg4\" (UID: \"99031ef1-ff62-4058-bf84-5c2240fbce40\") " pod="openstack/nova-cell1-db-create-s9sg4" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.628124 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5b99365-b407-4871-9930-1c7fa659c397-operator-scripts\") pod \"nova-cell0-db-create-8rt9f\" (UID: \"e5b99365-b407-4871-9930-1c7fa659c397\") " pod="openstack/nova-cell0-db-create-8rt9f" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.628165 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7a9cab4-be56-4a01-9b34-493370f1a56a-operator-scripts\") pod \"nova-api-7bf7-account-create-update-xqtj9\" (UID: \"f7a9cab4-be56-4a01-9b34-493370f1a56a\") " pod="openstack/nova-api-7bf7-account-create-update-xqtj9" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.628214 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgtnt\" (UniqueName: \"kubernetes.io/projected/f7a9cab4-be56-4a01-9b34-493370f1a56a-kube-api-access-kgtnt\") pod \"nova-api-7bf7-account-create-update-xqtj9\" (UID: \"f7a9cab4-be56-4a01-9b34-493370f1a56a\") " pod="openstack/nova-api-7bf7-account-create-update-xqtj9" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.628261 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh9m6\" (UniqueName: \"kubernetes.io/projected/99031ef1-ff62-4058-bf84-5c2240fbce40-kube-api-access-lh9m6\") pod \"nova-cell1-db-create-s9sg4\" (UID: \"99031ef1-ff62-4058-bf84-5c2240fbce40\") " pod="openstack/nova-cell1-db-create-s9sg4" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.628314 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e609dd6-5dee-4791-91e7-21e3191db678-operator-scripts\") pod \"nova-cell0-4825-account-create-update-k82lm\" (UID: \"0e609dd6-5dee-4791-91e7-21e3191db678\") " pod="openstack/nova-cell0-4825-account-create-update-k82lm" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.628333 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2v2r\" (UniqueName: \"kubernetes.io/projected/e5b99365-b407-4871-9930-1c7fa659c397-kube-api-access-f2v2r\") pod \"nova-cell0-db-create-8rt9f\" (UID: \"e5b99365-b407-4871-9930-1c7fa659c397\") " pod="openstack/nova-cell0-db-create-8rt9f" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.629058 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7a9cab4-be56-4a01-9b34-493370f1a56a-operator-scripts\") pod \"nova-api-7bf7-account-create-update-xqtj9\" (UID: \"f7a9cab4-be56-4a01-9b34-493370f1a56a\") " pod="openstack/nova-api-7bf7-account-create-update-xqtj9" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.629360 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5b99365-b407-4871-9930-1c7fa659c397-operator-scripts\") pod \"nova-cell0-db-create-8rt9f\" (UID: \"e5b99365-b407-4871-9930-1c7fa659c397\") " pod="openstack/nova-cell0-db-create-8rt9f" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.646873 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2v2r\" (UniqueName: \"kubernetes.io/projected/e5b99365-b407-4871-9930-1c7fa659c397-kube-api-access-f2v2r\") pod \"nova-cell0-db-create-8rt9f\" (UID: \"e5b99365-b407-4871-9930-1c7fa659c397\") " pod="openstack/nova-cell0-db-create-8rt9f" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.648151 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgtnt\" (UniqueName: \"kubernetes.io/projected/f7a9cab4-be56-4a01-9b34-493370f1a56a-kube-api-access-kgtnt\") pod \"nova-api-7bf7-account-create-update-xqtj9\" (UID: \"f7a9cab4-be56-4a01-9b34-493370f1a56a\") " pod="openstack/nova-api-7bf7-account-create-update-xqtj9" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.712856 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7bf7-account-create-update-xqtj9" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.729916 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh9m6\" (UniqueName: \"kubernetes.io/projected/99031ef1-ff62-4058-bf84-5c2240fbce40-kube-api-access-lh9m6\") pod \"nova-cell1-db-create-s9sg4\" (UID: \"99031ef1-ff62-4058-bf84-5c2240fbce40\") " pod="openstack/nova-cell1-db-create-s9sg4" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.730008 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e609dd6-5dee-4791-91e7-21e3191db678-operator-scripts\") pod \"nova-cell0-4825-account-create-update-k82lm\" (UID: \"0e609dd6-5dee-4791-91e7-21e3191db678\") " pod="openstack/nova-cell0-4825-account-create-update-k82lm" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.730084 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqvrb\" (UniqueName: \"kubernetes.io/projected/0e609dd6-5dee-4791-91e7-21e3191db678-kube-api-access-cqvrb\") pod \"nova-cell0-4825-account-create-update-k82lm\" (UID: \"0e609dd6-5dee-4791-91e7-21e3191db678\") " pod="openstack/nova-cell0-4825-account-create-update-k82lm" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.730634 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99031ef1-ff62-4058-bf84-5c2240fbce40-operator-scripts\") pod \"nova-cell1-db-create-s9sg4\" (UID: \"99031ef1-ff62-4058-bf84-5c2240fbce40\") " pod="openstack/nova-cell1-db-create-s9sg4" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.732538 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf" path="/var/lib/kubelet/pods/a76bc814-8c25-4caa-b7bd-cfcf8f8ecbbf/volumes" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.733004 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99031ef1-ff62-4058-bf84-5c2240fbce40-operator-scripts\") pod \"nova-cell1-db-create-s9sg4\" (UID: \"99031ef1-ff62-4058-bf84-5c2240fbce40\") " pod="openstack/nova-cell1-db-create-s9sg4" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.739733 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e609dd6-5dee-4791-91e7-21e3191db678-operator-scripts\") pod \"nova-cell0-4825-account-create-update-k82lm\" (UID: \"0e609dd6-5dee-4791-91e7-21e3191db678\") " pod="openstack/nova-cell0-4825-account-create-update-k82lm" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.740234 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-8rt9f" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.744383 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6ae06e4-ee1a-4519-beb5-23e71725f1a8" path="/var/lib/kubelet/pods/d6ae06e4-ee1a-4519-beb5-23e71725f1a8/volumes" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.753543 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh9m6\" (UniqueName: \"kubernetes.io/projected/99031ef1-ff62-4058-bf84-5c2240fbce40-kube-api-access-lh9m6\") pod \"nova-cell1-db-create-s9sg4\" (UID: \"99031ef1-ff62-4058-bf84-5c2240fbce40\") " pod="openstack/nova-cell1-db-create-s9sg4" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.763381 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqvrb\" (UniqueName: \"kubernetes.io/projected/0e609dd6-5dee-4791-91e7-21e3191db678-kube-api-access-cqvrb\") pod \"nova-cell0-4825-account-create-update-k82lm\" (UID: \"0e609dd6-5dee-4791-91e7-21e3191db678\") " pod="openstack/nova-cell0-4825-account-create-update-k82lm" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.764010 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-8cdb-account-create-update-xxcpw"] Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.765690 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8cdb-account-create-update-xxcpw" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.771984 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.776062 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-s9sg4" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.781045 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-8cdb-account-create-update-xxcpw"] Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.812297 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-4825-account-create-update-k82lm" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.934711 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0863e3d-3fe9-4f65-8436-6e2c8cc1165c-operator-scripts\") pod \"nova-cell1-8cdb-account-create-update-xxcpw\" (UID: \"c0863e3d-3fe9-4f65-8436-6e2c8cc1165c\") " pod="openstack/nova-cell1-8cdb-account-create-update-xxcpw" Dec 03 16:25:57 crc kubenswrapper[4998]: I1203 16:25:57.934811 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zlmj\" (UniqueName: \"kubernetes.io/projected/c0863e3d-3fe9-4f65-8436-6e2c8cc1165c-kube-api-access-9zlmj\") pod \"nova-cell1-8cdb-account-create-update-xxcpw\" (UID: \"c0863e3d-3fe9-4f65-8436-6e2c8cc1165c\") " pod="openstack/nova-cell1-8cdb-account-create-update-xxcpw" Dec 03 16:25:58 crc kubenswrapper[4998]: I1203 16:25:58.036597 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zlmj\" (UniqueName: \"kubernetes.io/projected/c0863e3d-3fe9-4f65-8436-6e2c8cc1165c-kube-api-access-9zlmj\") pod \"nova-cell1-8cdb-account-create-update-xxcpw\" (UID: \"c0863e3d-3fe9-4f65-8436-6e2c8cc1165c\") " pod="openstack/nova-cell1-8cdb-account-create-update-xxcpw" Dec 03 16:25:58 crc kubenswrapper[4998]: I1203 16:25:58.037080 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0863e3d-3fe9-4f65-8436-6e2c8cc1165c-operator-scripts\") pod \"nova-cell1-8cdb-account-create-update-xxcpw\" (UID: \"c0863e3d-3fe9-4f65-8436-6e2c8cc1165c\") " pod="openstack/nova-cell1-8cdb-account-create-update-xxcpw" Dec 03 16:25:58 crc kubenswrapper[4998]: I1203 16:25:58.037921 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0863e3d-3fe9-4f65-8436-6e2c8cc1165c-operator-scripts\") pod \"nova-cell1-8cdb-account-create-update-xxcpw\" (UID: \"c0863e3d-3fe9-4f65-8436-6e2c8cc1165c\") " pod="openstack/nova-cell1-8cdb-account-create-update-xxcpw" Dec 03 16:25:58 crc kubenswrapper[4998]: I1203 16:25:58.062392 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zlmj\" (UniqueName: \"kubernetes.io/projected/c0863e3d-3fe9-4f65-8436-6e2c8cc1165c-kube-api-access-9zlmj\") pod \"nova-cell1-8cdb-account-create-update-xxcpw\" (UID: \"c0863e3d-3fe9-4f65-8436-6e2c8cc1165c\") " pod="openstack/nova-cell1-8cdb-account-create-update-xxcpw" Dec 03 16:25:58 crc kubenswrapper[4998]: I1203 16:25:58.118256 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-tdm2m"] Dec 03 16:25:58 crc kubenswrapper[4998]: I1203 16:25:58.186830 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-7bf7-account-create-update-xqtj9"] Dec 03 16:25:58 crc kubenswrapper[4998]: W1203 16:25:58.226818 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7a9cab4_be56_4a01_9b34_493370f1a56a.slice/crio-133f6a6dd76768443b048b82a3fab155a0f2642e4a105f79b8f4f2d58631249f WatchSource:0}: Error finding container 133f6a6dd76768443b048b82a3fab155a0f2642e4a105f79b8f4f2d58631249f: Status 404 returned error can't find the container with id 133f6a6dd76768443b048b82a3fab155a0f2642e4a105f79b8f4f2d58631249f Dec 03 16:25:58 crc kubenswrapper[4998]: I1203 16:25:58.256556 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8cdb-account-create-update-xxcpw" Dec 03 16:25:58 crc kubenswrapper[4998]: I1203 16:25:58.281659 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb","Type":"ContainerStarted","Data":"16a67c27d69bc79ad920a839aa77b4f69a1111da0eb9383f8364f74c2627daa3"} Dec 03 16:25:58 crc kubenswrapper[4998]: I1203 16:25:58.286644 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"937de507-f2cb-4b8c-8b86-cfc929f23248","Type":"ContainerStarted","Data":"cafec0b9dcc43a250ef70c154996b1ca3cd1c464cdd95f9a1234ee8e44b9cfa1"} Dec 03 16:25:58 crc kubenswrapper[4998]: I1203 16:25:58.291054 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-tdm2m" event={"ID":"c49a0bb7-935a-4d43-bd6f-848d5e44b43f","Type":"ContainerStarted","Data":"76837c4b66d138cfcb1a0a1f95ad01f1e95688f461c7d4024cb909b203e31fe4"} Dec 03 16:25:58 crc kubenswrapper[4998]: I1203 16:25:58.525734 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-8rt9f"] Dec 03 16:25:58 crc kubenswrapper[4998]: I1203 16:25:58.585511 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-s9sg4"] Dec 03 16:25:58 crc kubenswrapper[4998]: I1203 16:25:58.598292 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-4825-account-create-update-k82lm"] Dec 03 16:25:58 crc kubenswrapper[4998]: W1203 16:25:58.604052 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99031ef1_ff62_4058_bf84_5c2240fbce40.slice/crio-a7b3f89145363c3cf21ad32e9b984bdcf4467b074d49a1cef3dd80f22169f1d1 WatchSource:0}: Error finding container a7b3f89145363c3cf21ad32e9b984bdcf4467b074d49a1cef3dd80f22169f1d1: Status 404 returned error can't find the container with id a7b3f89145363c3cf21ad32e9b984bdcf4467b074d49a1cef3dd80f22169f1d1 Dec 03 16:25:58 crc kubenswrapper[4998]: W1203 16:25:58.606461 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e609dd6_5dee_4791_91e7_21e3191db678.slice/crio-0c354b4518d6edce777080d85511bde9ab5e8d480461114b1c7e7fd84a51f0b9 WatchSource:0}: Error finding container 0c354b4518d6edce777080d85511bde9ab5e8d480461114b1c7e7fd84a51f0b9: Status 404 returned error can't find the container with id 0c354b4518d6edce777080d85511bde9ab5e8d480461114b1c7e7fd84a51f0b9 Dec 03 16:25:58 crc kubenswrapper[4998]: I1203 16:25:58.702638 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:25:58 crc kubenswrapper[4998]: I1203 16:25:58.708208 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7f47995564-5mql2" Dec 03 16:25:58 crc kubenswrapper[4998]: I1203 16:25:58.830044 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-8cdb-account-create-update-xxcpw"] Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.271219 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qjtcv"] Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.273199 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qjtcv" Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.286365 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qjtcv"] Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.338049 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-4825-account-create-update-k82lm" event={"ID":"0e609dd6-5dee-4791-91e7-21e3191db678","Type":"ContainerStarted","Data":"0c354b4518d6edce777080d85511bde9ab5e8d480461114b1c7e7fd84a51f0b9"} Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.359041 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb","Type":"ContainerStarted","Data":"086cd7a4103bcbadd1f82c39deab28fa7a86ed3cc5de8fd0f779b0d186f8f572"} Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.369268 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"937de507-f2cb-4b8c-8b86-cfc929f23248","Type":"ContainerStarted","Data":"c0ce97b555b61e9cc3e2278477bcf52e81b0905130ca158213b97689b4d7583e"} Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.373288 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a78fa691-c7a7-4908-8eb8-b3727b92941d-catalog-content\") pod \"redhat-operators-qjtcv\" (UID: \"a78fa691-c7a7-4908-8eb8-b3727b92941d\") " pod="openshift-marketplace/redhat-operators-qjtcv" Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.373366 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a78fa691-c7a7-4908-8eb8-b3727b92941d-utilities\") pod \"redhat-operators-qjtcv\" (UID: \"a78fa691-c7a7-4908-8eb8-b3727b92941d\") " pod="openshift-marketplace/redhat-operators-qjtcv" Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.373395 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wf84\" (UniqueName: \"kubernetes.io/projected/a78fa691-c7a7-4908-8eb8-b3727b92941d-kube-api-access-9wf84\") pod \"redhat-operators-qjtcv\" (UID: \"a78fa691-c7a7-4908-8eb8-b3727b92941d\") " pod="openshift-marketplace/redhat-operators-qjtcv" Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.381083 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8cdb-account-create-update-xxcpw" event={"ID":"c0863e3d-3fe9-4f65-8436-6e2c8cc1165c","Type":"ContainerStarted","Data":"3ae05721e5020d56ec12d8e0f9f14324a913b718288428997fb577a4e691b33a"} Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.384955 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7bf7-account-create-update-xqtj9" event={"ID":"f7a9cab4-be56-4a01-9b34-493370f1a56a","Type":"ContainerStarted","Data":"ce3e22e7f15f44efe3fb113b2f017f3130bb916c182926c239ef9495a5b26451"} Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.384988 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7bf7-account-create-update-xqtj9" event={"ID":"f7a9cab4-be56-4a01-9b34-493370f1a56a","Type":"ContainerStarted","Data":"133f6a6dd76768443b048b82a3fab155a0f2642e4a105f79b8f4f2d58631249f"} Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.389859 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-8rt9f" event={"ID":"e5b99365-b407-4871-9930-1c7fa659c397","Type":"ContainerStarted","Data":"ef79e83bb317586b636615f53ef93881cfbe790b6f3bb1a7c4e2d0776b1b189a"} Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.389890 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-8rt9f" event={"ID":"e5b99365-b407-4871-9930-1c7fa659c397","Type":"ContainerStarted","Data":"8cc6612027462b498cddc9c2491b72a7b346f4a755b22f1fc13904ecb358c96b"} Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.399784 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-tdm2m" event={"ID":"c49a0bb7-935a-4d43-bd6f-848d5e44b43f","Type":"ContainerStarted","Data":"aaf5b11bcc4c4e392a0d553686315edf17fa6e88837b90d37cec482da29159f7"} Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.407896 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-7bf7-account-create-update-xqtj9" podStartSLOduration=2.407880479 podStartE2EDuration="2.407880479s" podCreationTimestamp="2025-12-03 16:25:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:25:59.400854978 +0000 UTC m=+1338.012555201" watchObservedRunningTime="2025-12-03 16:25:59.407880479 +0000 UTC m=+1338.019580702" Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.426699 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-s9sg4" event={"ID":"99031ef1-ff62-4058-bf84-5c2240fbce40","Type":"ContainerStarted","Data":"2c4799dce3f2dcce67b1ade1fac3354b0087f5bc5469a7db466cc3741baad1d4"} Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.426742 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-s9sg4" event={"ID":"99031ef1-ff62-4058-bf84-5c2240fbce40","Type":"ContainerStarted","Data":"a7b3f89145363c3cf21ad32e9b984bdcf4467b074d49a1cef3dd80f22169f1d1"} Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.437950 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-tdm2m" podStartSLOduration=2.4379307199999998 podStartE2EDuration="2.43793072s" podCreationTimestamp="2025-12-03 16:25:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:25:59.415363211 +0000 UTC m=+1338.027063434" watchObservedRunningTime="2025-12-03 16:25:59.43793072 +0000 UTC m=+1338.049630943" Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.464575 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-s9sg4" podStartSLOduration=2.464557627 podStartE2EDuration="2.464557627s" podCreationTimestamp="2025-12-03 16:25:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:25:59.446849207 +0000 UTC m=+1338.058549430" watchObservedRunningTime="2025-12-03 16:25:59.464557627 +0000 UTC m=+1338.076257850" Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.477382 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a78fa691-c7a7-4908-8eb8-b3727b92941d-catalog-content\") pod \"redhat-operators-qjtcv\" (UID: \"a78fa691-c7a7-4908-8eb8-b3727b92941d\") " pod="openshift-marketplace/redhat-operators-qjtcv" Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.477487 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a78fa691-c7a7-4908-8eb8-b3727b92941d-utilities\") pod \"redhat-operators-qjtcv\" (UID: \"a78fa691-c7a7-4908-8eb8-b3727b92941d\") " pod="openshift-marketplace/redhat-operators-qjtcv" Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.477517 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wf84\" (UniqueName: \"kubernetes.io/projected/a78fa691-c7a7-4908-8eb8-b3727b92941d-kube-api-access-9wf84\") pod \"redhat-operators-qjtcv\" (UID: \"a78fa691-c7a7-4908-8eb8-b3727b92941d\") " pod="openshift-marketplace/redhat-operators-qjtcv" Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.478646 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a78fa691-c7a7-4908-8eb8-b3727b92941d-catalog-content\") pod \"redhat-operators-qjtcv\" (UID: \"a78fa691-c7a7-4908-8eb8-b3727b92941d\") " pod="openshift-marketplace/redhat-operators-qjtcv" Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.479513 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a78fa691-c7a7-4908-8eb8-b3727b92941d-utilities\") pod \"redhat-operators-qjtcv\" (UID: \"a78fa691-c7a7-4908-8eb8-b3727b92941d\") " pod="openshift-marketplace/redhat-operators-qjtcv" Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.504833 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wf84\" (UniqueName: \"kubernetes.io/projected/a78fa691-c7a7-4908-8eb8-b3727b92941d-kube-api-access-9wf84\") pod \"redhat-operators-qjtcv\" (UID: \"a78fa691-c7a7-4908-8eb8-b3727b92941d\") " pod="openshift-marketplace/redhat-operators-qjtcv" Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.515528 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.515859 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/watcher-decision-engine-0" Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.516424 4998 scope.go:117] "RemoveContainer" containerID="b657ed22e9096f43b7242cbed330ab7de4b3e3c8d87141fb5d8555f0560c2457" Dec 03 16:25:59 crc kubenswrapper[4998]: I1203 16:25:59.594036 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qjtcv" Dec 03 16:26:00 crc kubenswrapper[4998]: I1203 16:26:00.111280 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qjtcv"] Dec 03 16:26:00 crc kubenswrapper[4998]: W1203 16:26:00.155979 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda78fa691_c7a7_4908_8eb8_b3727b92941d.slice/crio-edc42995ad9c0e08eeb54a10a5dd7ecaec8751942eea70f64e501ccd9c4730c1 WatchSource:0}: Error finding container edc42995ad9c0e08eeb54a10a5dd7ecaec8751942eea70f64e501ccd9c4730c1: Status 404 returned error can't find the container with id edc42995ad9c0e08eeb54a10a5dd7ecaec8751942eea70f64e501ccd9c4730c1 Dec 03 16:26:00 crc kubenswrapper[4998]: I1203 16:26:00.437919 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"a009aeee-c849-4c00-b41b-3cf5e36a75bb","Type":"ContainerStarted","Data":"c923ec83ef3968e5e7f59971085d59629092a44767625d2597d9b5e9e3e8177e"} Dec 03 16:26:00 crc kubenswrapper[4998]: I1203 16:26:00.442569 4998 generic.go:334] "Generic (PLEG): container finished" podID="f7a9cab4-be56-4a01-9b34-493370f1a56a" containerID="ce3e22e7f15f44efe3fb113b2f017f3130bb916c182926c239ef9495a5b26451" exitCode=0 Dec 03 16:26:00 crc kubenswrapper[4998]: I1203 16:26:00.442635 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7bf7-account-create-update-xqtj9" event={"ID":"f7a9cab4-be56-4a01-9b34-493370f1a56a","Type":"ContainerDied","Data":"ce3e22e7f15f44efe3fb113b2f017f3130bb916c182926c239ef9495a5b26451"} Dec 03 16:26:00 crc kubenswrapper[4998]: I1203 16:26:00.444484 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb","Type":"ContainerStarted","Data":"90d18d7b758db275df459fb07a945bf60dfb96c97ce23119c055e7e1350b7352"} Dec 03 16:26:00 crc kubenswrapper[4998]: I1203 16:26:00.447016 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qjtcv" event={"ID":"a78fa691-c7a7-4908-8eb8-b3727b92941d","Type":"ContainerStarted","Data":"edc42995ad9c0e08eeb54a10a5dd7ecaec8751942eea70f64e501ccd9c4730c1"} Dec 03 16:26:00 crc kubenswrapper[4998]: I1203 16:26:00.448369 4998 generic.go:334] "Generic (PLEG): container finished" podID="c0863e3d-3fe9-4f65-8436-6e2c8cc1165c" containerID="65a19393bd94e0171b263597878f2a7b35cfe646f0319e3410af84795657581a" exitCode=0 Dec 03 16:26:00 crc kubenswrapper[4998]: I1203 16:26:00.448417 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8cdb-account-create-update-xxcpw" event={"ID":"c0863e3d-3fe9-4f65-8436-6e2c8cc1165c","Type":"ContainerDied","Data":"65a19393bd94e0171b263597878f2a7b35cfe646f0319e3410af84795657581a"} Dec 03 16:26:00 crc kubenswrapper[4998]: I1203 16:26:00.449980 4998 generic.go:334] "Generic (PLEG): container finished" podID="e5b99365-b407-4871-9930-1c7fa659c397" containerID="ef79e83bb317586b636615f53ef93881cfbe790b6f3bb1a7c4e2d0776b1b189a" exitCode=0 Dec 03 16:26:00 crc kubenswrapper[4998]: I1203 16:26:00.450018 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-8rt9f" event={"ID":"e5b99365-b407-4871-9930-1c7fa659c397","Type":"ContainerDied","Data":"ef79e83bb317586b636615f53ef93881cfbe790b6f3bb1a7c4e2d0776b1b189a"} Dec 03 16:26:00 crc kubenswrapper[4998]: I1203 16:26:00.451217 4998 generic.go:334] "Generic (PLEG): container finished" podID="c49a0bb7-935a-4d43-bd6f-848d5e44b43f" containerID="aaf5b11bcc4c4e392a0d553686315edf17fa6e88837b90d37cec482da29159f7" exitCode=0 Dec 03 16:26:00 crc kubenswrapper[4998]: I1203 16:26:00.451254 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-tdm2m" event={"ID":"c49a0bb7-935a-4d43-bd6f-848d5e44b43f","Type":"ContainerDied","Data":"aaf5b11bcc4c4e392a0d553686315edf17fa6e88837b90d37cec482da29159f7"} Dec 03 16:26:00 crc kubenswrapper[4998]: I1203 16:26:00.452582 4998 generic.go:334] "Generic (PLEG): container finished" podID="99031ef1-ff62-4058-bf84-5c2240fbce40" containerID="2c4799dce3f2dcce67b1ade1fac3354b0087f5bc5469a7db466cc3741baad1d4" exitCode=0 Dec 03 16:26:00 crc kubenswrapper[4998]: I1203 16:26:00.452625 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-s9sg4" event={"ID":"99031ef1-ff62-4058-bf84-5c2240fbce40","Type":"ContainerDied","Data":"2c4799dce3f2dcce67b1ade1fac3354b0087f5bc5469a7db466cc3741baad1d4"} Dec 03 16:26:00 crc kubenswrapper[4998]: I1203 16:26:00.456298 4998 generic.go:334] "Generic (PLEG): container finished" podID="0e609dd6-5dee-4791-91e7-21e3191db678" containerID="523211a913430c4a44b7d89513310d008be76421fe3f439806c255ab5d8d8db6" exitCode=0 Dec 03 16:26:00 crc kubenswrapper[4998]: I1203 16:26:00.456342 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-4825-account-create-update-k82lm" event={"ID":"0e609dd6-5dee-4791-91e7-21e3191db678","Type":"ContainerDied","Data":"523211a913430c4a44b7d89513310d008be76421fe3f439806c255ab5d8d8db6"} Dec 03 16:26:01 crc kubenswrapper[4998]: I1203 16:26:01.051170 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-77c59f659f-jkw4x" Dec 03 16:26:01 crc kubenswrapper[4998]: I1203 16:26:01.126028 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7f47995564-5mql2"] Dec 03 16:26:01 crc kubenswrapper[4998]: I1203 16:26:01.126241 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7f47995564-5mql2" podUID="98bb0f8f-aa52-4408-8d16-c6677b080101" containerName="neutron-api" containerID="cri-o://5ab8347b81ea9f11541eee83ec6cd743c0c89c981fcef844dbf98fb32b6c3599" gracePeriod=30 Dec 03 16:26:01 crc kubenswrapper[4998]: I1203 16:26:01.126547 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7f47995564-5mql2" podUID="98bb0f8f-aa52-4408-8d16-c6677b080101" containerName="neutron-httpd" containerID="cri-o://a1d30b67a1dfea698ecb9b1bea63314a38f9de7898fc55974c57c35c76092c39" gracePeriod=30 Dec 03 16:26:01 crc kubenswrapper[4998]: I1203 16:26:01.475867 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"937de507-f2cb-4b8c-8b86-cfc929f23248","Type":"ContainerStarted","Data":"48a79a62d7a2061f0aae43efabb741d527b4aa33f939c21d7e412d6261f50a86"} Dec 03 16:26:01 crc kubenswrapper[4998]: I1203 16:26:01.476000 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 16:26:01 crc kubenswrapper[4998]: I1203 16:26:01.480396 4998 generic.go:334] "Generic (PLEG): container finished" podID="a78fa691-c7a7-4908-8eb8-b3727b92941d" containerID="863de87c4eb155ccc524703ea1a7fb18595e6f92612c8969ca41517db5289aa5" exitCode=0 Dec 03 16:26:01 crc kubenswrapper[4998]: I1203 16:26:01.481180 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qjtcv" event={"ID":"a78fa691-c7a7-4908-8eb8-b3727b92941d","Type":"ContainerDied","Data":"863de87c4eb155ccc524703ea1a7fb18595e6f92612c8969ca41517db5289aa5"} Dec 03 16:26:01 crc kubenswrapper[4998]: I1203 16:26:01.515526 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.515504027 podStartE2EDuration="5.515504027s" podCreationTimestamp="2025-12-03 16:25:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:26:01.501033375 +0000 UTC m=+1340.112733598" watchObservedRunningTime="2025-12-03 16:26:01.515504027 +0000 UTC m=+1340.127204250" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.010311 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-s9sg4" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.137869 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lh9m6\" (UniqueName: \"kubernetes.io/projected/99031ef1-ff62-4058-bf84-5c2240fbce40-kube-api-access-lh9m6\") pod \"99031ef1-ff62-4058-bf84-5c2240fbce40\" (UID: \"99031ef1-ff62-4058-bf84-5c2240fbce40\") " Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.138132 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99031ef1-ff62-4058-bf84-5c2240fbce40-operator-scripts\") pod \"99031ef1-ff62-4058-bf84-5c2240fbce40\" (UID: \"99031ef1-ff62-4058-bf84-5c2240fbce40\") " Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.139247 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99031ef1-ff62-4058-bf84-5c2240fbce40-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "99031ef1-ff62-4058-bf84-5c2240fbce40" (UID: "99031ef1-ff62-4058-bf84-5c2240fbce40"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.142954 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99031ef1-ff62-4058-bf84-5c2240fbce40-kube-api-access-lh9m6" (OuterVolumeSpecName: "kube-api-access-lh9m6") pod "99031ef1-ff62-4058-bf84-5c2240fbce40" (UID: "99031ef1-ff62-4058-bf84-5c2240fbce40"). InnerVolumeSpecName "kube-api-access-lh9m6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.227618 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-8rt9f" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.245492 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lh9m6\" (UniqueName: \"kubernetes.io/projected/99031ef1-ff62-4058-bf84-5c2240fbce40-kube-api-access-lh9m6\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.245524 4998 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99031ef1-ff62-4058-bf84-5c2240fbce40-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.252780 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8cdb-account-create-update-xxcpw" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.262038 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-4825-account-create-update-k82lm" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.273984 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-tdm2m" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.291654 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7bf7-account-create-update-xqtj9" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.346267 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zlmj\" (UniqueName: \"kubernetes.io/projected/c0863e3d-3fe9-4f65-8436-6e2c8cc1165c-kube-api-access-9zlmj\") pod \"c0863e3d-3fe9-4f65-8436-6e2c8cc1165c\" (UID: \"c0863e3d-3fe9-4f65-8436-6e2c8cc1165c\") " Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.346328 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqvrb\" (UniqueName: \"kubernetes.io/projected/0e609dd6-5dee-4791-91e7-21e3191db678-kube-api-access-cqvrb\") pod \"0e609dd6-5dee-4791-91e7-21e3191db678\" (UID: \"0e609dd6-5dee-4791-91e7-21e3191db678\") " Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.346450 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5b99365-b407-4871-9930-1c7fa659c397-operator-scripts\") pod \"e5b99365-b407-4871-9930-1c7fa659c397\" (UID: \"e5b99365-b407-4871-9930-1c7fa659c397\") " Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.346501 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0863e3d-3fe9-4f65-8436-6e2c8cc1165c-operator-scripts\") pod \"c0863e3d-3fe9-4f65-8436-6e2c8cc1165c\" (UID: \"c0863e3d-3fe9-4f65-8436-6e2c8cc1165c\") " Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.346533 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c49a0bb7-935a-4d43-bd6f-848d5e44b43f-operator-scripts\") pod \"c49a0bb7-935a-4d43-bd6f-848d5e44b43f\" (UID: \"c49a0bb7-935a-4d43-bd6f-848d5e44b43f\") " Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.346584 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7a9cab4-be56-4a01-9b34-493370f1a56a-operator-scripts\") pod \"f7a9cab4-be56-4a01-9b34-493370f1a56a\" (UID: \"f7a9cab4-be56-4a01-9b34-493370f1a56a\") " Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.346623 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2v2r\" (UniqueName: \"kubernetes.io/projected/e5b99365-b407-4871-9930-1c7fa659c397-kube-api-access-f2v2r\") pod \"e5b99365-b407-4871-9930-1c7fa659c397\" (UID: \"e5b99365-b407-4871-9930-1c7fa659c397\") " Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.346686 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsh2t\" (UniqueName: \"kubernetes.io/projected/c49a0bb7-935a-4d43-bd6f-848d5e44b43f-kube-api-access-zsh2t\") pod \"c49a0bb7-935a-4d43-bd6f-848d5e44b43f\" (UID: \"c49a0bb7-935a-4d43-bd6f-848d5e44b43f\") " Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.346784 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgtnt\" (UniqueName: \"kubernetes.io/projected/f7a9cab4-be56-4a01-9b34-493370f1a56a-kube-api-access-kgtnt\") pod \"f7a9cab4-be56-4a01-9b34-493370f1a56a\" (UID: \"f7a9cab4-be56-4a01-9b34-493370f1a56a\") " Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.346811 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e609dd6-5dee-4791-91e7-21e3191db678-operator-scripts\") pod \"0e609dd6-5dee-4791-91e7-21e3191db678\" (UID: \"0e609dd6-5dee-4791-91e7-21e3191db678\") " Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.346993 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0863e3d-3fe9-4f65-8436-6e2c8cc1165c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c0863e3d-3fe9-4f65-8436-6e2c8cc1165c" (UID: "c0863e3d-3fe9-4f65-8436-6e2c8cc1165c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.347315 4998 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0863e3d-3fe9-4f65-8436-6e2c8cc1165c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.347686 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5b99365-b407-4871-9930-1c7fa659c397-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e5b99365-b407-4871-9930-1c7fa659c397" (UID: "e5b99365-b407-4871-9930-1c7fa659c397"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.347713 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e609dd6-5dee-4791-91e7-21e3191db678-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0e609dd6-5dee-4791-91e7-21e3191db678" (UID: "0e609dd6-5dee-4791-91e7-21e3191db678"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.348164 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7a9cab4-be56-4a01-9b34-493370f1a56a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f7a9cab4-be56-4a01-9b34-493370f1a56a" (UID: "f7a9cab4-be56-4a01-9b34-493370f1a56a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.349803 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0863e3d-3fe9-4f65-8436-6e2c8cc1165c-kube-api-access-9zlmj" (OuterVolumeSpecName: "kube-api-access-9zlmj") pod "c0863e3d-3fe9-4f65-8436-6e2c8cc1165c" (UID: "c0863e3d-3fe9-4f65-8436-6e2c8cc1165c"). InnerVolumeSpecName "kube-api-access-9zlmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.350110 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c49a0bb7-935a-4d43-bd6f-848d5e44b43f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c49a0bb7-935a-4d43-bd6f-848d5e44b43f" (UID: "c49a0bb7-935a-4d43-bd6f-848d5e44b43f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.351684 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5b99365-b407-4871-9930-1c7fa659c397-kube-api-access-f2v2r" (OuterVolumeSpecName: "kube-api-access-f2v2r") pod "e5b99365-b407-4871-9930-1c7fa659c397" (UID: "e5b99365-b407-4871-9930-1c7fa659c397"). InnerVolumeSpecName "kube-api-access-f2v2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.355994 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7a9cab4-be56-4a01-9b34-493370f1a56a-kube-api-access-kgtnt" (OuterVolumeSpecName: "kube-api-access-kgtnt") pod "f7a9cab4-be56-4a01-9b34-493370f1a56a" (UID: "f7a9cab4-be56-4a01-9b34-493370f1a56a"). InnerVolumeSpecName "kube-api-access-kgtnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.356049 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c49a0bb7-935a-4d43-bd6f-848d5e44b43f-kube-api-access-zsh2t" (OuterVolumeSpecName: "kube-api-access-zsh2t") pod "c49a0bb7-935a-4d43-bd6f-848d5e44b43f" (UID: "c49a0bb7-935a-4d43-bd6f-848d5e44b43f"). InnerVolumeSpecName "kube-api-access-zsh2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.356068 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e609dd6-5dee-4791-91e7-21e3191db678-kube-api-access-cqvrb" (OuterVolumeSpecName: "kube-api-access-cqvrb") pod "0e609dd6-5dee-4791-91e7-21e3191db678" (UID: "0e609dd6-5dee-4791-91e7-21e3191db678"). InnerVolumeSpecName "kube-api-access-cqvrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.448569 4998 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5b99365-b407-4871-9930-1c7fa659c397-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.448598 4998 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c49a0bb7-935a-4d43-bd6f-848d5e44b43f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.448609 4998 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7a9cab4-be56-4a01-9b34-493370f1a56a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.448618 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2v2r\" (UniqueName: \"kubernetes.io/projected/e5b99365-b407-4871-9930-1c7fa659c397-kube-api-access-f2v2r\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.448629 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsh2t\" (UniqueName: \"kubernetes.io/projected/c49a0bb7-935a-4d43-bd6f-848d5e44b43f-kube-api-access-zsh2t\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.448637 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgtnt\" (UniqueName: \"kubernetes.io/projected/f7a9cab4-be56-4a01-9b34-493370f1a56a-kube-api-access-kgtnt\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.448647 4998 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e609dd6-5dee-4791-91e7-21e3191db678-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.448656 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zlmj\" (UniqueName: \"kubernetes.io/projected/c0863e3d-3fe9-4f65-8436-6e2c8cc1165c-kube-api-access-9zlmj\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.448665 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqvrb\" (UniqueName: \"kubernetes.io/projected/0e609dd6-5dee-4791-91e7-21e3191db678-kube-api-access-cqvrb\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.495322 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-tdm2m" event={"ID":"c49a0bb7-935a-4d43-bd6f-848d5e44b43f","Type":"ContainerDied","Data":"76837c4b66d138cfcb1a0a1f95ad01f1e95688f461c7d4024cb909b203e31fe4"} Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.495361 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76837c4b66d138cfcb1a0a1f95ad01f1e95688f461c7d4024cb909b203e31fe4" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.495414 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-tdm2m" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.504268 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.504495 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="1d7de5e1-fc35-4b0c-8b1d-509c30d11939" containerName="glance-log" containerID="cri-o://89c33d9ec7f919bebf6bcd6a1e9679473e0531814045827c4bca1cb28a037a00" gracePeriod=30 Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.504896 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="1d7de5e1-fc35-4b0c-8b1d-509c30d11939" containerName="glance-httpd" containerID="cri-o://4e719b9f4645604bbdc077703eb3389a2f670a438a0887e292ab1a908019f135" gracePeriod=30 Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.508878 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-s9sg4" event={"ID":"99031ef1-ff62-4058-bf84-5c2240fbce40","Type":"ContainerDied","Data":"a7b3f89145363c3cf21ad32e9b984bdcf4467b074d49a1cef3dd80f22169f1d1"} Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.508916 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7b3f89145363c3cf21ad32e9b984bdcf4467b074d49a1cef3dd80f22169f1d1" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.509020 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-s9sg4" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.516672 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-8rt9f" event={"ID":"e5b99365-b407-4871-9930-1c7fa659c397","Type":"ContainerDied","Data":"8cc6612027462b498cddc9c2491b72a7b346f4a755b22f1fc13904ecb358c96b"} Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.516730 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8cc6612027462b498cddc9c2491b72a7b346f4a755b22f1fc13904ecb358c96b" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.516821 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-8rt9f" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.519135 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-4825-account-create-update-k82lm" event={"ID":"0e609dd6-5dee-4791-91e7-21e3191db678","Type":"ContainerDied","Data":"0c354b4518d6edce777080d85511bde9ab5e8d480461114b1c7e7fd84a51f0b9"} Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.519170 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c354b4518d6edce777080d85511bde9ab5e8d480461114b1c7e7fd84a51f0b9" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.519220 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-4825-account-create-update-k82lm" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.538484 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb","Type":"ContainerStarted","Data":"4844ae774e7f67fadde0dc4fd9801b6efc10b9ace28a14f79732af4985e41603"} Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.538767 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.538787 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" containerName="ceilometer-central-agent" containerID="cri-o://16a67c27d69bc79ad920a839aa77b4f69a1111da0eb9383f8364f74c2627daa3" gracePeriod=30 Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.538842 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" containerName="proxy-httpd" containerID="cri-o://4844ae774e7f67fadde0dc4fd9801b6efc10b9ace28a14f79732af4985e41603" gracePeriod=30 Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.538897 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" containerName="sg-core" containerID="cri-o://90d18d7b758db275df459fb07a945bf60dfb96c97ce23119c055e7e1350b7352" gracePeriod=30 Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.538988 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" containerName="ceilometer-notification-agent" containerID="cri-o://086cd7a4103bcbadd1f82c39deab28fa7a86ed3cc5de8fd0f779b0d186f8f572" gracePeriod=30 Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.566232 4998 generic.go:334] "Generic (PLEG): container finished" podID="98bb0f8f-aa52-4408-8d16-c6677b080101" containerID="a1d30b67a1dfea698ecb9b1bea63314a38f9de7898fc55974c57c35c76092c39" exitCode=0 Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.566328 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f47995564-5mql2" event={"ID":"98bb0f8f-aa52-4408-8d16-c6677b080101","Type":"ContainerDied","Data":"a1d30b67a1dfea698ecb9b1bea63314a38f9de7898fc55974c57c35c76092c39"} Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.584213 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.042739521 podStartE2EDuration="6.584191084s" podCreationTimestamp="2025-12-03 16:25:56 +0000 UTC" firstStartedPulling="2025-12-03 16:25:57.169356395 +0000 UTC m=+1335.781056618" lastFinishedPulling="2025-12-03 16:26:01.710807958 +0000 UTC m=+1340.322508181" observedRunningTime="2025-12-03 16:26:02.566430352 +0000 UTC m=+1341.178130575" watchObservedRunningTime="2025-12-03 16:26:02.584191084 +0000 UTC m=+1341.195891307" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.619305 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qjtcv" event={"ID":"a78fa691-c7a7-4908-8eb8-b3727b92941d","Type":"ContainerStarted","Data":"f5fda72cc17b1098cc27f616651a1083ad5dc310101edb27f5380804017543e6"} Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.642345 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-8cdb-account-create-update-xxcpw" event={"ID":"c0863e3d-3fe9-4f65-8436-6e2c8cc1165c","Type":"ContainerDied","Data":"3ae05721e5020d56ec12d8e0f9f14324a913b718288428997fb577a4e691b33a"} Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.642404 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ae05721e5020d56ec12d8e0f9f14324a913b718288428997fb577a4e691b33a" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.642500 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-8cdb-account-create-update-xxcpw" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.650531 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7bf7-account-create-update-xqtj9" Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.650566 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7bf7-account-create-update-xqtj9" event={"ID":"f7a9cab4-be56-4a01-9b34-493370f1a56a","Type":"ContainerDied","Data":"133f6a6dd76768443b048b82a3fab155a0f2642e4a105f79b8f4f2d58631249f"} Dec 03 16:26:02 crc kubenswrapper[4998]: I1203 16:26:02.650588 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="133f6a6dd76768443b048b82a3fab155a0f2642e4a105f79b8f4f2d58631249f" Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.663563 4998 generic.go:334] "Generic (PLEG): container finished" podID="1d7de5e1-fc35-4b0c-8b1d-509c30d11939" containerID="4e719b9f4645604bbdc077703eb3389a2f670a438a0887e292ab1a908019f135" exitCode=0 Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.663600 4998 generic.go:334] "Generic (PLEG): container finished" podID="1d7de5e1-fc35-4b0c-8b1d-509c30d11939" containerID="89c33d9ec7f919bebf6bcd6a1e9679473e0531814045827c4bca1cb28a037a00" exitCode=143 Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.663643 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1d7de5e1-fc35-4b0c-8b1d-509c30d11939","Type":"ContainerDied","Data":"4e719b9f4645604bbdc077703eb3389a2f670a438a0887e292ab1a908019f135"} Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.663693 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1d7de5e1-fc35-4b0c-8b1d-509c30d11939","Type":"ContainerDied","Data":"89c33d9ec7f919bebf6bcd6a1e9679473e0531814045827c4bca1cb28a037a00"} Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.666491 4998 generic.go:334] "Generic (PLEG): container finished" podID="a78fa691-c7a7-4908-8eb8-b3727b92941d" containerID="f5fda72cc17b1098cc27f616651a1083ad5dc310101edb27f5380804017543e6" exitCode=0 Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.666558 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qjtcv" event={"ID":"a78fa691-c7a7-4908-8eb8-b3727b92941d","Type":"ContainerDied","Data":"f5fda72cc17b1098cc27f616651a1083ad5dc310101edb27f5380804017543e6"} Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.692448 4998 generic.go:334] "Generic (PLEG): container finished" podID="3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" containerID="4844ae774e7f67fadde0dc4fd9801b6efc10b9ace28a14f79732af4985e41603" exitCode=0 Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.692768 4998 generic.go:334] "Generic (PLEG): container finished" podID="3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" containerID="90d18d7b758db275df459fb07a945bf60dfb96c97ce23119c055e7e1350b7352" exitCode=2 Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.692781 4998 generic.go:334] "Generic (PLEG): container finished" podID="3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" containerID="086cd7a4103bcbadd1f82c39deab28fa7a86ed3cc5de8fd0f779b0d186f8f572" exitCode=0 Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.713060 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb","Type":"ContainerDied","Data":"4844ae774e7f67fadde0dc4fd9801b6efc10b9ace28a14f79732af4985e41603"} Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.713102 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb","Type":"ContainerDied","Data":"90d18d7b758db275df459fb07a945bf60dfb96c97ce23119c055e7e1350b7352"} Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.713115 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb","Type":"ContainerDied","Data":"086cd7a4103bcbadd1f82c39deab28fa7a86ed3cc5de8fd0f779b0d186f8f572"} Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.903665 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.984191 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-logs\") pod \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.984245 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-httpd-run\") pod \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.984281 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.984331 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-public-tls-certs\") pod \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.984466 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-combined-ca-bundle\") pod \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.984496 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8z96p\" (UniqueName: \"kubernetes.io/projected/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-kube-api-access-8z96p\") pod \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.984539 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-scripts\") pod \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.984688 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-config-data\") pod \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\" (UID: \"1d7de5e1-fc35-4b0c-8b1d-509c30d11939\") " Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.984881 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1d7de5e1-fc35-4b0c-8b1d-509c30d11939" (UID: "1d7de5e1-fc35-4b0c-8b1d-509c30d11939"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.984991 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-logs" (OuterVolumeSpecName: "logs") pod "1d7de5e1-fc35-4b0c-8b1d-509c30d11939" (UID: "1d7de5e1-fc35-4b0c-8b1d-509c30d11939"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.985460 4998 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-logs\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.985480 4998 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:03 crc kubenswrapper[4998]: I1203 16:26:03.991435 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "1d7de5e1-fc35-4b0c-8b1d-509c30d11939" (UID: "1d7de5e1-fc35-4b0c-8b1d-509c30d11939"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.006959 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-kube-api-access-8z96p" (OuterVolumeSpecName: "kube-api-access-8z96p") pod "1d7de5e1-fc35-4b0c-8b1d-509c30d11939" (UID: "1d7de5e1-fc35-4b0c-8b1d-509c30d11939"). InnerVolumeSpecName "kube-api-access-8z96p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.009729 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-scripts" (OuterVolumeSpecName: "scripts") pod "1d7de5e1-fc35-4b0c-8b1d-509c30d11939" (UID: "1d7de5e1-fc35-4b0c-8b1d-509c30d11939"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.071019 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-config-data" (OuterVolumeSpecName: "config-data") pod "1d7de5e1-fc35-4b0c-8b1d-509c30d11939" (UID: "1d7de5e1-fc35-4b0c-8b1d-509c30d11939"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.073991 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1d7de5e1-fc35-4b0c-8b1d-509c30d11939" (UID: "1d7de5e1-fc35-4b0c-8b1d-509c30d11939"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.084948 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d7de5e1-fc35-4b0c-8b1d-509c30d11939" (UID: "1d7de5e1-fc35-4b0c-8b1d-509c30d11939"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.087054 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.087102 4998 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.087112 4998 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.087122 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.087131 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8z96p\" (UniqueName: \"kubernetes.io/projected/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-kube-api-access-8z96p\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.087142 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d7de5e1-fc35-4b0c-8b1d-509c30d11939-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.127609 4998 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.155626 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.155897 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d4832eb3-4e67-43a9-abc0-a8b9680ea7d4" containerName="glance-log" containerID="cri-o://1657b6f33a5e6e5d2f68ab2fa8bc506670d5df5ca49603145704a3b5d0967132" gracePeriod=30 Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.156000 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d4832eb3-4e67-43a9-abc0-a8b9680ea7d4" containerName="glance-httpd" containerID="cri-o://9b5be307c8ff3af1a80772f8810dddbc54affcdf58a7fb498209aac41357a527" gracePeriod=30 Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.199108 4998 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.710333 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.710547 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1d7de5e1-fc35-4b0c-8b1d-509c30d11939","Type":"ContainerDied","Data":"db56ab4a8d6a9bf092191e5b57c7957937d3e81cd7afb128942a1e1ce461120a"} Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.710603 4998 scope.go:117] "RemoveContainer" containerID="4e719b9f4645604bbdc077703eb3389a2f670a438a0887e292ab1a908019f135" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.721219 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qjtcv" event={"ID":"a78fa691-c7a7-4908-8eb8-b3727b92941d","Type":"ContainerStarted","Data":"d90119e30812730816317a3ca8fc16f0f405cd659a5eaee1babf027e432b911d"} Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.727729 4998 generic.go:334] "Generic (PLEG): container finished" podID="d4832eb3-4e67-43a9-abc0-a8b9680ea7d4" containerID="1657b6f33a5e6e5d2f68ab2fa8bc506670d5df5ca49603145704a3b5d0967132" exitCode=143 Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.727831 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4","Type":"ContainerDied","Data":"1657b6f33a5e6e5d2f68ab2fa8bc506670d5df5ca49603145704a3b5d0967132"} Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.746429 4998 scope.go:117] "RemoveContainer" containerID="89c33d9ec7f919bebf6bcd6a1e9679473e0531814045827c4bca1cb28a037a00" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.768972 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qjtcv" podStartSLOduration=3.170623474 podStartE2EDuration="5.768951939s" podCreationTimestamp="2025-12-03 16:25:59 +0000 UTC" firstStartedPulling="2025-12-03 16:26:01.485823185 +0000 UTC m=+1340.097523408" lastFinishedPulling="2025-12-03 16:26:04.08415165 +0000 UTC m=+1342.695851873" observedRunningTime="2025-12-03 16:26:04.742131316 +0000 UTC m=+1343.353831539" watchObservedRunningTime="2025-12-03 16:26:04.768951939 +0000 UTC m=+1343.380652162" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.822648 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.841684 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.849738 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 16:26:04 crc kubenswrapper[4998]: E1203 16:26:04.850218 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d7de5e1-fc35-4b0c-8b1d-509c30d11939" containerName="glance-httpd" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.850239 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d7de5e1-fc35-4b0c-8b1d-509c30d11939" containerName="glance-httpd" Dec 03 16:26:04 crc kubenswrapper[4998]: E1203 16:26:04.850256 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5b99365-b407-4871-9930-1c7fa659c397" containerName="mariadb-database-create" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.850263 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5b99365-b407-4871-9930-1c7fa659c397" containerName="mariadb-database-create" Dec 03 16:26:04 crc kubenswrapper[4998]: E1203 16:26:04.850275 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7a9cab4-be56-4a01-9b34-493370f1a56a" containerName="mariadb-account-create-update" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.850281 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7a9cab4-be56-4a01-9b34-493370f1a56a" containerName="mariadb-account-create-update" Dec 03 16:26:04 crc kubenswrapper[4998]: E1203 16:26:04.850292 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c49a0bb7-935a-4d43-bd6f-848d5e44b43f" containerName="mariadb-database-create" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.850298 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="c49a0bb7-935a-4d43-bd6f-848d5e44b43f" containerName="mariadb-database-create" Dec 03 16:26:04 crc kubenswrapper[4998]: E1203 16:26:04.850312 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99031ef1-ff62-4058-bf84-5c2240fbce40" containerName="mariadb-database-create" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.850318 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="99031ef1-ff62-4058-bf84-5c2240fbce40" containerName="mariadb-database-create" Dec 03 16:26:04 crc kubenswrapper[4998]: E1203 16:26:04.850339 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0863e3d-3fe9-4f65-8436-6e2c8cc1165c" containerName="mariadb-account-create-update" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.850345 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0863e3d-3fe9-4f65-8436-6e2c8cc1165c" containerName="mariadb-account-create-update" Dec 03 16:26:04 crc kubenswrapper[4998]: E1203 16:26:04.850359 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d7de5e1-fc35-4b0c-8b1d-509c30d11939" containerName="glance-log" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.850365 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d7de5e1-fc35-4b0c-8b1d-509c30d11939" containerName="glance-log" Dec 03 16:26:04 crc kubenswrapper[4998]: E1203 16:26:04.850373 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e609dd6-5dee-4791-91e7-21e3191db678" containerName="mariadb-account-create-update" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.850378 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e609dd6-5dee-4791-91e7-21e3191db678" containerName="mariadb-account-create-update" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.850543 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7a9cab4-be56-4a01-9b34-493370f1a56a" containerName="mariadb-account-create-update" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.850552 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d7de5e1-fc35-4b0c-8b1d-509c30d11939" containerName="glance-httpd" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.850565 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d7de5e1-fc35-4b0c-8b1d-509c30d11939" containerName="glance-log" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.850572 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0863e3d-3fe9-4f65-8436-6e2c8cc1165c" containerName="mariadb-account-create-update" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.850585 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="99031ef1-ff62-4058-bf84-5c2240fbce40" containerName="mariadb-database-create" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.850598 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e609dd6-5dee-4791-91e7-21e3191db678" containerName="mariadb-account-create-update" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.850607 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="c49a0bb7-935a-4d43-bd6f-848d5e44b43f" containerName="mariadb-database-create" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.850622 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5b99365-b407-4871-9930-1c7fa659c397" containerName="mariadb-database-create" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.852399 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.856122 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.856481 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.866565 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.919882 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l955t\" (UniqueName: \"kubernetes.io/projected/1cbdcfac-d2fe-456c-9632-f67aca0d05d5-kube-api-access-l955t\") pod \"glance-default-external-api-0\" (UID: \"1cbdcfac-d2fe-456c-9632-f67aca0d05d5\") " pod="openstack/glance-default-external-api-0" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.920015 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cbdcfac-d2fe-456c-9632-f67aca0d05d5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1cbdcfac-d2fe-456c-9632-f67aca0d05d5\") " pod="openstack/glance-default-external-api-0" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.920077 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1cbdcfac-d2fe-456c-9632-f67aca0d05d5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1cbdcfac-d2fe-456c-9632-f67aca0d05d5\") " pod="openstack/glance-default-external-api-0" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.920254 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cbdcfac-d2fe-456c-9632-f67aca0d05d5-config-data\") pod \"glance-default-external-api-0\" (UID: \"1cbdcfac-d2fe-456c-9632-f67aca0d05d5\") " pod="openstack/glance-default-external-api-0" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.920343 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cbdcfac-d2fe-456c-9632-f67aca0d05d5-scripts\") pod \"glance-default-external-api-0\" (UID: \"1cbdcfac-d2fe-456c-9632-f67aca0d05d5\") " pod="openstack/glance-default-external-api-0" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.920435 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1cbdcfac-d2fe-456c-9632-f67aca0d05d5-logs\") pod \"glance-default-external-api-0\" (UID: \"1cbdcfac-d2fe-456c-9632-f67aca0d05d5\") " pod="openstack/glance-default-external-api-0" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.920644 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"1cbdcfac-d2fe-456c-9632-f67aca0d05d5\") " pod="openstack/glance-default-external-api-0" Dec 03 16:26:04 crc kubenswrapper[4998]: I1203 16:26:04.920807 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1cbdcfac-d2fe-456c-9632-f67aca0d05d5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1cbdcfac-d2fe-456c-9632-f67aca0d05d5\") " pod="openstack/glance-default-external-api-0" Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.022918 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"1cbdcfac-d2fe-456c-9632-f67aca0d05d5\") " pod="openstack/glance-default-external-api-0" Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.023002 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1cbdcfac-d2fe-456c-9632-f67aca0d05d5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1cbdcfac-d2fe-456c-9632-f67aca0d05d5\") " pod="openstack/glance-default-external-api-0" Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.023050 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l955t\" (UniqueName: \"kubernetes.io/projected/1cbdcfac-d2fe-456c-9632-f67aca0d05d5-kube-api-access-l955t\") pod \"glance-default-external-api-0\" (UID: \"1cbdcfac-d2fe-456c-9632-f67aca0d05d5\") " pod="openstack/glance-default-external-api-0" Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.023089 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cbdcfac-d2fe-456c-9632-f67aca0d05d5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1cbdcfac-d2fe-456c-9632-f67aca0d05d5\") " pod="openstack/glance-default-external-api-0" Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.023118 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1cbdcfac-d2fe-456c-9632-f67aca0d05d5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1cbdcfac-d2fe-456c-9632-f67aca0d05d5\") " pod="openstack/glance-default-external-api-0" Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.023163 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cbdcfac-d2fe-456c-9632-f67aca0d05d5-config-data\") pod \"glance-default-external-api-0\" (UID: \"1cbdcfac-d2fe-456c-9632-f67aca0d05d5\") " pod="openstack/glance-default-external-api-0" Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.023196 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cbdcfac-d2fe-456c-9632-f67aca0d05d5-scripts\") pod \"glance-default-external-api-0\" (UID: \"1cbdcfac-d2fe-456c-9632-f67aca0d05d5\") " pod="openstack/glance-default-external-api-0" Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.023243 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1cbdcfac-d2fe-456c-9632-f67aca0d05d5-logs\") pod \"glance-default-external-api-0\" (UID: \"1cbdcfac-d2fe-456c-9632-f67aca0d05d5\") " pod="openstack/glance-default-external-api-0" Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.023463 4998 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"1cbdcfac-d2fe-456c-9632-f67aca0d05d5\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.023809 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1cbdcfac-d2fe-456c-9632-f67aca0d05d5-logs\") pod \"glance-default-external-api-0\" (UID: \"1cbdcfac-d2fe-456c-9632-f67aca0d05d5\") " pod="openstack/glance-default-external-api-0" Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.024080 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1cbdcfac-d2fe-456c-9632-f67aca0d05d5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"1cbdcfac-d2fe-456c-9632-f67aca0d05d5\") " pod="openstack/glance-default-external-api-0" Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.034655 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1cbdcfac-d2fe-456c-9632-f67aca0d05d5-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"1cbdcfac-d2fe-456c-9632-f67aca0d05d5\") " pod="openstack/glance-default-external-api-0" Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.042290 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cbdcfac-d2fe-456c-9632-f67aca0d05d5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"1cbdcfac-d2fe-456c-9632-f67aca0d05d5\") " pod="openstack/glance-default-external-api-0" Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.042503 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1cbdcfac-d2fe-456c-9632-f67aca0d05d5-config-data\") pod \"glance-default-external-api-0\" (UID: \"1cbdcfac-d2fe-456c-9632-f67aca0d05d5\") " pod="openstack/glance-default-external-api-0" Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.047341 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1cbdcfac-d2fe-456c-9632-f67aca0d05d5-scripts\") pod \"glance-default-external-api-0\" (UID: \"1cbdcfac-d2fe-456c-9632-f67aca0d05d5\") " pod="openstack/glance-default-external-api-0" Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.048879 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l955t\" (UniqueName: \"kubernetes.io/projected/1cbdcfac-d2fe-456c-9632-f67aca0d05d5-kube-api-access-l955t\") pod \"glance-default-external-api-0\" (UID: \"1cbdcfac-d2fe-456c-9632-f67aca0d05d5\") " pod="openstack/glance-default-external-api-0" Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.091216 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"1cbdcfac-d2fe-456c-9632-f67aca0d05d5\") " pod="openstack/glance-default-external-api-0" Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.170104 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.694672 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d7de5e1-fc35-4b0c-8b1d-509c30d11939" path="/var/lib/kubelet/pods/1d7de5e1-fc35-4b0c-8b1d-509c30d11939/volumes" Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.747628 4998 generic.go:334] "Generic (PLEG): container finished" podID="98bb0f8f-aa52-4408-8d16-c6677b080101" containerID="5ab8347b81ea9f11541eee83ec6cd743c0c89c981fcef844dbf98fb32b6c3599" exitCode=0 Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.747716 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f47995564-5mql2" event={"ID":"98bb0f8f-aa52-4408-8d16-c6677b080101","Type":"ContainerDied","Data":"5ab8347b81ea9f11541eee83ec6cd743c0c89c981fcef844dbf98fb32b6c3599"} Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.747748 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7f47995564-5mql2" event={"ID":"98bb0f8f-aa52-4408-8d16-c6677b080101","Type":"ContainerDied","Data":"c458dd983344d1d0d82c52ca5b5caabfef781971f092857283c28e64df5fbe35"} Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.747774 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c458dd983344d1d0d82c52ca5b5caabfef781971f092857283c28e64df5fbe35" Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.784014 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 16:26:05 crc kubenswrapper[4998]: W1203 16:26:05.831474 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1cbdcfac_d2fe_456c_9632_f67aca0d05d5.slice/crio-280aa664162af1b88d836c15bd52106e407990d097cd4045c431643b2387c5c5 WatchSource:0}: Error finding container 280aa664162af1b88d836c15bd52106e407990d097cd4045c431643b2387c5c5: Status 404 returned error can't find the container with id 280aa664162af1b88d836c15bd52106e407990d097cd4045c431643b2387c5c5 Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.865921 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7f47995564-5mql2" Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.941087 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsmw6\" (UniqueName: \"kubernetes.io/projected/98bb0f8f-aa52-4408-8d16-c6677b080101-kube-api-access-tsmw6\") pod \"98bb0f8f-aa52-4408-8d16-c6677b080101\" (UID: \"98bb0f8f-aa52-4408-8d16-c6677b080101\") " Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.941230 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/98bb0f8f-aa52-4408-8d16-c6677b080101-ovndb-tls-certs\") pod \"98bb0f8f-aa52-4408-8d16-c6677b080101\" (UID: \"98bb0f8f-aa52-4408-8d16-c6677b080101\") " Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.941275 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98bb0f8f-aa52-4408-8d16-c6677b080101-combined-ca-bundle\") pod \"98bb0f8f-aa52-4408-8d16-c6677b080101\" (UID: \"98bb0f8f-aa52-4408-8d16-c6677b080101\") " Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.941331 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/98bb0f8f-aa52-4408-8d16-c6677b080101-httpd-config\") pod \"98bb0f8f-aa52-4408-8d16-c6677b080101\" (UID: \"98bb0f8f-aa52-4408-8d16-c6677b080101\") " Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.941410 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/98bb0f8f-aa52-4408-8d16-c6677b080101-config\") pod \"98bb0f8f-aa52-4408-8d16-c6677b080101\" (UID: \"98bb0f8f-aa52-4408-8d16-c6677b080101\") " Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.953752 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98bb0f8f-aa52-4408-8d16-c6677b080101-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "98bb0f8f-aa52-4408-8d16-c6677b080101" (UID: "98bb0f8f-aa52-4408-8d16-c6677b080101"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:05 crc kubenswrapper[4998]: I1203 16:26:05.954082 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98bb0f8f-aa52-4408-8d16-c6677b080101-kube-api-access-tsmw6" (OuterVolumeSpecName: "kube-api-access-tsmw6") pod "98bb0f8f-aa52-4408-8d16-c6677b080101" (UID: "98bb0f8f-aa52-4408-8d16-c6677b080101"). InnerVolumeSpecName "kube-api-access-tsmw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.008441 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98bb0f8f-aa52-4408-8d16-c6677b080101-config" (OuterVolumeSpecName: "config") pod "98bb0f8f-aa52-4408-8d16-c6677b080101" (UID: "98bb0f8f-aa52-4408-8d16-c6677b080101"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.013069 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98bb0f8f-aa52-4408-8d16-c6677b080101-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "98bb0f8f-aa52-4408-8d16-c6677b080101" (UID: "98bb0f8f-aa52-4408-8d16-c6677b080101"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.043147 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98bb0f8f-aa52-4408-8d16-c6677b080101-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "98bb0f8f-aa52-4408-8d16-c6677b080101" (UID: "98bb0f8f-aa52-4408-8d16-c6677b080101"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.043546 4998 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/98bb0f8f-aa52-4408-8d16-c6677b080101-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.043576 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/98bb0f8f-aa52-4408-8d16-c6677b080101-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.043586 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsmw6\" (UniqueName: \"kubernetes.io/projected/98bb0f8f-aa52-4408-8d16-c6677b080101-kube-api-access-tsmw6\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.043599 4998 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/98bb0f8f-aa52-4408-8d16-c6677b080101-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.043607 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98bb0f8f-aa52-4408-8d16-c6677b080101-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.269089 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.351271 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkf7v\" (UniqueName: \"kubernetes.io/projected/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-kube-api-access-fkf7v\") pod \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.351457 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-combined-ca-bundle\") pod \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.351933 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.352022 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-httpd-run\") pod \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.352060 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-config-data\") pod \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.352154 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-logs\") pod \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.352186 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-scripts\") pod \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.352217 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-internal-tls-certs\") pod \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\" (UID: \"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4\") " Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.354205 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-logs" (OuterVolumeSpecName: "logs") pod "d4832eb3-4e67-43a9-abc0-a8b9680ea7d4" (UID: "d4832eb3-4e67-43a9-abc0-a8b9680ea7d4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.354246 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d4832eb3-4e67-43a9-abc0-a8b9680ea7d4" (UID: "d4832eb3-4e67-43a9-abc0-a8b9680ea7d4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.357943 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "d4832eb3-4e67-43a9-abc0-a8b9680ea7d4" (UID: "d4832eb3-4e67-43a9-abc0-a8b9680ea7d4"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.358478 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-scripts" (OuterVolumeSpecName: "scripts") pod "d4832eb3-4e67-43a9-abc0-a8b9680ea7d4" (UID: "d4832eb3-4e67-43a9-abc0-a8b9680ea7d4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.377951 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-kube-api-access-fkf7v" (OuterVolumeSpecName: "kube-api-access-fkf7v") pod "d4832eb3-4e67-43a9-abc0-a8b9680ea7d4" (UID: "d4832eb3-4e67-43a9-abc0-a8b9680ea7d4"). InnerVolumeSpecName "kube-api-access-fkf7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.432923 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4832eb3-4e67-43a9-abc0-a8b9680ea7d4" (UID: "d4832eb3-4e67-43a9-abc0-a8b9680ea7d4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.437947 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d4832eb3-4e67-43a9-abc0-a8b9680ea7d4" (UID: "d4832eb3-4e67-43a9-abc0-a8b9680ea7d4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.452351 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-config-data" (OuterVolumeSpecName: "config-data") pod "d4832eb3-4e67-43a9-abc0-a8b9680ea7d4" (UID: "d4832eb3-4e67-43a9-abc0-a8b9680ea7d4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.455012 4998 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-logs\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.455041 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.455054 4998 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.455068 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkf7v\" (UniqueName: \"kubernetes.io/projected/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-kube-api-access-fkf7v\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.455080 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.455109 4998 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.455122 4998 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.455134 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.495383 4998 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.557409 4998 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.765421 4998 generic.go:334] "Generic (PLEG): container finished" podID="d4832eb3-4e67-43a9-abc0-a8b9680ea7d4" containerID="9b5be307c8ff3af1a80772f8810dddbc54affcdf58a7fb498209aac41357a527" exitCode=0 Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.765488 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4","Type":"ContainerDied","Data":"9b5be307c8ff3af1a80772f8810dddbc54affcdf58a7fb498209aac41357a527"} Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.765514 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d4832eb3-4e67-43a9-abc0-a8b9680ea7d4","Type":"ContainerDied","Data":"585b6021eb6b0fe8eefab2d6ddbb6af3eba05c312be8c6d32d8926b631ed91ca"} Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.765532 4998 scope.go:117] "RemoveContainer" containerID="9b5be307c8ff3af1a80772f8810dddbc54affcdf58a7fb498209aac41357a527" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.765624 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.772233 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7f47995564-5mql2" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.772901 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1cbdcfac-d2fe-456c-9632-f67aca0d05d5","Type":"ContainerStarted","Data":"280aa664162af1b88d836c15bd52106e407990d097cd4045c431643b2387c5c5"} Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.796463 4998 scope.go:117] "RemoveContainer" containerID="1657b6f33a5e6e5d2f68ab2fa8bc506670d5df5ca49603145704a3b5d0967132" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.806820 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.823857 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.850474 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 16:26:06 crc kubenswrapper[4998]: E1203 16:26:06.850894 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98bb0f8f-aa52-4408-8d16-c6677b080101" containerName="neutron-api" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.850906 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="98bb0f8f-aa52-4408-8d16-c6677b080101" containerName="neutron-api" Dec 03 16:26:06 crc kubenswrapper[4998]: E1203 16:26:06.850917 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98bb0f8f-aa52-4408-8d16-c6677b080101" containerName="neutron-httpd" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.850923 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="98bb0f8f-aa52-4408-8d16-c6677b080101" containerName="neutron-httpd" Dec 03 16:26:06 crc kubenswrapper[4998]: E1203 16:26:06.850944 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4832eb3-4e67-43a9-abc0-a8b9680ea7d4" containerName="glance-log" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.850950 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4832eb3-4e67-43a9-abc0-a8b9680ea7d4" containerName="glance-log" Dec 03 16:26:06 crc kubenswrapper[4998]: E1203 16:26:06.850965 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4832eb3-4e67-43a9-abc0-a8b9680ea7d4" containerName="glance-httpd" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.850970 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4832eb3-4e67-43a9-abc0-a8b9680ea7d4" containerName="glance-httpd" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.851159 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="98bb0f8f-aa52-4408-8d16-c6677b080101" containerName="neutron-api" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.851170 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4832eb3-4e67-43a9-abc0-a8b9680ea7d4" containerName="glance-log" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.851195 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="98bb0f8f-aa52-4408-8d16-c6677b080101" containerName="neutron-httpd" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.851207 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4832eb3-4e67-43a9-abc0-a8b9680ea7d4" containerName="glance-httpd" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.852218 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.862592 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.871672 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.887015 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.974715 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"7874dd07-6286-4514-8896-309a007a9aee\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.975053 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7874dd07-6286-4514-8896-309a007a9aee-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7874dd07-6286-4514-8896-309a007a9aee\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.975098 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjf4p\" (UniqueName: \"kubernetes.io/projected/7874dd07-6286-4514-8896-309a007a9aee-kube-api-access-hjf4p\") pod \"glance-default-internal-api-0\" (UID: \"7874dd07-6286-4514-8896-309a007a9aee\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.975134 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7874dd07-6286-4514-8896-309a007a9aee-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7874dd07-6286-4514-8896-309a007a9aee\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.975188 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7874dd07-6286-4514-8896-309a007a9aee-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7874dd07-6286-4514-8896-309a007a9aee\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.975247 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7874dd07-6286-4514-8896-309a007a9aee-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7874dd07-6286-4514-8896-309a007a9aee\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.975277 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7874dd07-6286-4514-8896-309a007a9aee-logs\") pod \"glance-default-internal-api-0\" (UID: \"7874dd07-6286-4514-8896-309a007a9aee\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.975561 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7874dd07-6286-4514-8896-309a007a9aee-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7874dd07-6286-4514-8896-309a007a9aee\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.989466 4998 scope.go:117] "RemoveContainer" containerID="9b5be307c8ff3af1a80772f8810dddbc54affcdf58a7fb498209aac41357a527" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.990381 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7f47995564-5mql2"] Dec 03 16:26:06 crc kubenswrapper[4998]: E1203 16:26:06.992579 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b5be307c8ff3af1a80772f8810dddbc54affcdf58a7fb498209aac41357a527\": container with ID starting with 9b5be307c8ff3af1a80772f8810dddbc54affcdf58a7fb498209aac41357a527 not found: ID does not exist" containerID="9b5be307c8ff3af1a80772f8810dddbc54affcdf58a7fb498209aac41357a527" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.992617 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b5be307c8ff3af1a80772f8810dddbc54affcdf58a7fb498209aac41357a527"} err="failed to get container status \"9b5be307c8ff3af1a80772f8810dddbc54affcdf58a7fb498209aac41357a527\": rpc error: code = NotFound desc = could not find container \"9b5be307c8ff3af1a80772f8810dddbc54affcdf58a7fb498209aac41357a527\": container with ID starting with 9b5be307c8ff3af1a80772f8810dddbc54affcdf58a7fb498209aac41357a527 not found: ID does not exist" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.992643 4998 scope.go:117] "RemoveContainer" containerID="1657b6f33a5e6e5d2f68ab2fa8bc506670d5df5ca49603145704a3b5d0967132" Dec 03 16:26:06 crc kubenswrapper[4998]: E1203 16:26:06.992910 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1657b6f33a5e6e5d2f68ab2fa8bc506670d5df5ca49603145704a3b5d0967132\": container with ID starting with 1657b6f33a5e6e5d2f68ab2fa8bc506670d5df5ca49603145704a3b5d0967132 not found: ID does not exist" containerID="1657b6f33a5e6e5d2f68ab2fa8bc506670d5df5ca49603145704a3b5d0967132" Dec 03 16:26:06 crc kubenswrapper[4998]: I1203 16:26:06.992938 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1657b6f33a5e6e5d2f68ab2fa8bc506670d5df5ca49603145704a3b5d0967132"} err="failed to get container status \"1657b6f33a5e6e5d2f68ab2fa8bc506670d5df5ca49603145704a3b5d0967132\": rpc error: code = NotFound desc = could not find container \"1657b6f33a5e6e5d2f68ab2fa8bc506670d5df5ca49603145704a3b5d0967132\": container with ID starting with 1657b6f33a5e6e5d2f68ab2fa8bc506670d5df5ca49603145704a3b5d0967132 not found: ID does not exist" Dec 03 16:26:07 crc kubenswrapper[4998]: I1203 16:26:07.014819 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7f47995564-5mql2"] Dec 03 16:26:07 crc kubenswrapper[4998]: I1203 16:26:07.082932 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7874dd07-6286-4514-8896-309a007a9aee-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7874dd07-6286-4514-8896-309a007a9aee\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:26:07 crc kubenswrapper[4998]: I1203 16:26:07.083077 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"7874dd07-6286-4514-8896-309a007a9aee\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:26:07 crc kubenswrapper[4998]: I1203 16:26:07.083105 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7874dd07-6286-4514-8896-309a007a9aee-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7874dd07-6286-4514-8896-309a007a9aee\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:26:07 crc kubenswrapper[4998]: I1203 16:26:07.083138 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjf4p\" (UniqueName: \"kubernetes.io/projected/7874dd07-6286-4514-8896-309a007a9aee-kube-api-access-hjf4p\") pod \"glance-default-internal-api-0\" (UID: \"7874dd07-6286-4514-8896-309a007a9aee\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:26:07 crc kubenswrapper[4998]: I1203 16:26:07.083188 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7874dd07-6286-4514-8896-309a007a9aee-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7874dd07-6286-4514-8896-309a007a9aee\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:26:07 crc kubenswrapper[4998]: I1203 16:26:07.083217 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7874dd07-6286-4514-8896-309a007a9aee-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7874dd07-6286-4514-8896-309a007a9aee\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:26:07 crc kubenswrapper[4998]: I1203 16:26:07.083250 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7874dd07-6286-4514-8896-309a007a9aee-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7874dd07-6286-4514-8896-309a007a9aee\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:26:07 crc kubenswrapper[4998]: I1203 16:26:07.083278 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7874dd07-6286-4514-8896-309a007a9aee-logs\") pod \"glance-default-internal-api-0\" (UID: \"7874dd07-6286-4514-8896-309a007a9aee\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:26:07 crc kubenswrapper[4998]: I1203 16:26:07.083838 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7874dd07-6286-4514-8896-309a007a9aee-logs\") pod \"glance-default-internal-api-0\" (UID: \"7874dd07-6286-4514-8896-309a007a9aee\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:26:07 crc kubenswrapper[4998]: I1203 16:26:07.085458 4998 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"7874dd07-6286-4514-8896-309a007a9aee\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Dec 03 16:26:07 crc kubenswrapper[4998]: I1203 16:26:07.087041 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7874dd07-6286-4514-8896-309a007a9aee-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7874dd07-6286-4514-8896-309a007a9aee\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:26:07 crc kubenswrapper[4998]: I1203 16:26:07.090978 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7874dd07-6286-4514-8896-309a007a9aee-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7874dd07-6286-4514-8896-309a007a9aee\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:26:07 crc kubenswrapper[4998]: I1203 16:26:07.094378 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7874dd07-6286-4514-8896-309a007a9aee-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7874dd07-6286-4514-8896-309a007a9aee\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:26:07 crc kubenswrapper[4998]: I1203 16:26:07.096909 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7874dd07-6286-4514-8896-309a007a9aee-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7874dd07-6286-4514-8896-309a007a9aee\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:26:07 crc kubenswrapper[4998]: I1203 16:26:07.097662 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7874dd07-6286-4514-8896-309a007a9aee-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7874dd07-6286-4514-8896-309a007a9aee\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:26:07 crc kubenswrapper[4998]: I1203 16:26:07.104717 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjf4p\" (UniqueName: \"kubernetes.io/projected/7874dd07-6286-4514-8896-309a007a9aee-kube-api-access-hjf4p\") pod \"glance-default-internal-api-0\" (UID: \"7874dd07-6286-4514-8896-309a007a9aee\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:26:07 crc kubenswrapper[4998]: I1203 16:26:07.131434 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"7874dd07-6286-4514-8896-309a007a9aee\") " pod="openstack/glance-default-internal-api-0" Dec 03 16:26:07 crc kubenswrapper[4998]: I1203 16:26:07.273722 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 16:26:07 crc kubenswrapper[4998]: I1203 16:26:07.692117 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98bb0f8f-aa52-4408-8d16-c6677b080101" path="/var/lib/kubelet/pods/98bb0f8f-aa52-4408-8d16-c6677b080101/volumes" Dec 03 16:26:07 crc kubenswrapper[4998]: I1203 16:26:07.693119 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4832eb3-4e67-43a9-abc0-a8b9680ea7d4" path="/var/lib/kubelet/pods/d4832eb3-4e67-43a9-abc0-a8b9680ea7d4/volumes" Dec 03 16:26:07 crc kubenswrapper[4998]: I1203 16:26:07.783415 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1cbdcfac-d2fe-456c-9632-f67aca0d05d5","Type":"ContainerStarted","Data":"b211ec4c06e9944cf57be1536185abf1b1fa5d7badb2ed11ed1f5b9854280a09"} Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.017268 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6zkx5"] Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.018608 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6zkx5" Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.021447 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.021676 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-xcs6z" Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.021860 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.030526 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6zkx5"] Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.082638 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.103483 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ff4ba4e-9b43-41fd-902b-977b9cc3aba9-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-6zkx5\" (UID: \"3ff4ba4e-9b43-41fd-902b-977b9cc3aba9\") " pod="openstack/nova-cell0-conductor-db-sync-6zkx5" Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.103569 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lbmx\" (UniqueName: \"kubernetes.io/projected/3ff4ba4e-9b43-41fd-902b-977b9cc3aba9-kube-api-access-4lbmx\") pod \"nova-cell0-conductor-db-sync-6zkx5\" (UID: \"3ff4ba4e-9b43-41fd-902b-977b9cc3aba9\") " pod="openstack/nova-cell0-conductor-db-sync-6zkx5" Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.103639 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ff4ba4e-9b43-41fd-902b-977b9cc3aba9-scripts\") pod \"nova-cell0-conductor-db-sync-6zkx5\" (UID: \"3ff4ba4e-9b43-41fd-902b-977b9cc3aba9\") " pod="openstack/nova-cell0-conductor-db-sync-6zkx5" Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.104167 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ff4ba4e-9b43-41fd-902b-977b9cc3aba9-config-data\") pod \"nova-cell0-conductor-db-sync-6zkx5\" (UID: \"3ff4ba4e-9b43-41fd-902b-977b9cc3aba9\") " pod="openstack/nova-cell0-conductor-db-sync-6zkx5" Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.206286 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ff4ba4e-9b43-41fd-902b-977b9cc3aba9-config-data\") pod \"nova-cell0-conductor-db-sync-6zkx5\" (UID: \"3ff4ba4e-9b43-41fd-902b-977b9cc3aba9\") " pod="openstack/nova-cell0-conductor-db-sync-6zkx5" Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.206697 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ff4ba4e-9b43-41fd-902b-977b9cc3aba9-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-6zkx5\" (UID: \"3ff4ba4e-9b43-41fd-902b-977b9cc3aba9\") " pod="openstack/nova-cell0-conductor-db-sync-6zkx5" Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.206792 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lbmx\" (UniqueName: \"kubernetes.io/projected/3ff4ba4e-9b43-41fd-902b-977b9cc3aba9-kube-api-access-4lbmx\") pod \"nova-cell0-conductor-db-sync-6zkx5\" (UID: \"3ff4ba4e-9b43-41fd-902b-977b9cc3aba9\") " pod="openstack/nova-cell0-conductor-db-sync-6zkx5" Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.206883 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ff4ba4e-9b43-41fd-902b-977b9cc3aba9-scripts\") pod \"nova-cell0-conductor-db-sync-6zkx5\" (UID: \"3ff4ba4e-9b43-41fd-902b-977b9cc3aba9\") " pod="openstack/nova-cell0-conductor-db-sync-6zkx5" Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.218174 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ff4ba4e-9b43-41fd-902b-977b9cc3aba9-scripts\") pod \"nova-cell0-conductor-db-sync-6zkx5\" (UID: \"3ff4ba4e-9b43-41fd-902b-977b9cc3aba9\") " pod="openstack/nova-cell0-conductor-db-sync-6zkx5" Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.218287 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ff4ba4e-9b43-41fd-902b-977b9cc3aba9-config-data\") pod \"nova-cell0-conductor-db-sync-6zkx5\" (UID: \"3ff4ba4e-9b43-41fd-902b-977b9cc3aba9\") " pod="openstack/nova-cell0-conductor-db-sync-6zkx5" Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.218903 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ff4ba4e-9b43-41fd-902b-977b9cc3aba9-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-6zkx5\" (UID: \"3ff4ba4e-9b43-41fd-902b-977b9cc3aba9\") " pod="openstack/nova-cell0-conductor-db-sync-6zkx5" Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.232145 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lbmx\" (UniqueName: \"kubernetes.io/projected/3ff4ba4e-9b43-41fd-902b-977b9cc3aba9-kube-api-access-4lbmx\") pod \"nova-cell0-conductor-db-sync-6zkx5\" (UID: \"3ff4ba4e-9b43-41fd-902b-977b9cc3aba9\") " pod="openstack/nova-cell0-conductor-db-sync-6zkx5" Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.348391 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6zkx5" Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.810612 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6zkx5"] Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.813582 4998 generic.go:334] "Generic (PLEG): container finished" podID="3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" containerID="16a67c27d69bc79ad920a839aa77b4f69a1111da0eb9383f8364f74c2627daa3" exitCode=0 Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.813643 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb","Type":"ContainerDied","Data":"16a67c27d69bc79ad920a839aa77b4f69a1111da0eb9383f8364f74c2627daa3"} Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.815263 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7874dd07-6286-4514-8896-309a007a9aee","Type":"ContainerStarted","Data":"70367ad7b3525497cd88cdd0fa0436248498c58eed0f908915bc67285f761775"} Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.815286 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7874dd07-6286-4514-8896-309a007a9aee","Type":"ContainerStarted","Data":"23f4ce0cfaa4e147037f6d91ac0471d2554d2865b7e7ce92ccd8ea678c313801"} Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.816489 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"1cbdcfac-d2fe-456c-9632-f67aca0d05d5","Type":"ContainerStarted","Data":"76392161636fb56e8b06c71070f7977f4a3d24c1d447ee95e1638e88c2a97fa2"} Dec 03 16:26:08 crc kubenswrapper[4998]: I1203 16:26:08.853605 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.853582769 podStartE2EDuration="4.853582769s" podCreationTimestamp="2025-12-03 16:26:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:26:08.845970134 +0000 UTC m=+1347.457670357" watchObservedRunningTime="2025-12-03 16:26:08.853582769 +0000 UTC m=+1347.465282992" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.108293 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.130979 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrc4z\" (UniqueName: \"kubernetes.io/projected/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-kube-api-access-jrc4z\") pod \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.131068 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-config-data\") pod \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.131096 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-log-httpd\") pod \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.131113 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-run-httpd\") pod \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.131172 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-sg-core-conf-yaml\") pod \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.131228 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-combined-ca-bundle\") pod \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.131329 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-scripts\") pod \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\" (UID: \"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb\") " Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.132404 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" (UID: "3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.132429 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" (UID: "3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.139019 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-kube-api-access-jrc4z" (OuterVolumeSpecName: "kube-api-access-jrc4z") pod "3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" (UID: "3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb"). InnerVolumeSpecName "kube-api-access-jrc4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.147955 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-scripts" (OuterVolumeSpecName: "scripts") pod "3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" (UID: "3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.210947 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" (UID: "3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.235953 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.235997 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrc4z\" (UniqueName: \"kubernetes.io/projected/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-kube-api-access-jrc4z\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.236013 4998 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.236025 4998 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.236036 4998 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.287368 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" (UID: "3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.300041 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-config-data" (OuterVolumeSpecName: "config-data") pod "3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" (UID: "3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.338694 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.338938 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.514331 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.561504 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.604020 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qjtcv" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.604065 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qjtcv" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.671891 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.835092 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6zkx5" event={"ID":"3ff4ba4e-9b43-41fd-902b-977b9cc3aba9","Type":"ContainerStarted","Data":"2f4fd4359a332b21964064a20bc7de44fd4fd1822fdefccedc0cb1dab703e357"} Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.836987 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb","Type":"ContainerDied","Data":"9d17b2b0f53adc3a709aa486133f615dfde872e9ac0e2a248cb5c799030d5856"} Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.837023 4998 scope.go:117] "RemoveContainer" containerID="4844ae774e7f67fadde0dc4fd9801b6efc10b9ace28a14f79732af4985e41603" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.837165 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.842603 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7874dd07-6286-4514-8896-309a007a9aee","Type":"ContainerStarted","Data":"7a530a10d19d886538c999464c8b463f5140a3fc164ed4806245399a58e73641"} Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.843102 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.865804 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.875058 4998 scope.go:117] "RemoveContainer" containerID="90d18d7b758db275df459fb07a945bf60dfb96c97ce23119c055e7e1350b7352" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.883913 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.887036 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.925661 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:26:09 crc kubenswrapper[4998]: E1203 16:26:09.926135 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" containerName="proxy-httpd" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.926153 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" containerName="proxy-httpd" Dec 03 16:26:09 crc kubenswrapper[4998]: E1203 16:26:09.926164 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" containerName="ceilometer-central-agent" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.926170 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" containerName="ceilometer-central-agent" Dec 03 16:26:09 crc kubenswrapper[4998]: E1203 16:26:09.926180 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" containerName="sg-core" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.926186 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" containerName="sg-core" Dec 03 16:26:09 crc kubenswrapper[4998]: E1203 16:26:09.926200 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" containerName="ceilometer-notification-agent" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.926208 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" containerName="ceilometer-notification-agent" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.926379 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" containerName="ceilometer-notification-agent" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.926402 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" containerName="proxy-httpd" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.926412 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" containerName="sg-core" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.926424 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" containerName="ceilometer-central-agent" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.928099 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.934657 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.934933 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.944897 4998 scope.go:117] "RemoveContainer" containerID="086cd7a4103bcbadd1f82c39deab28fa7a86ed3cc5de8fd0f779b0d186f8f572" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.947122 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7hb7\" (UniqueName: \"kubernetes.io/projected/ee4992d5-6a42-46d6-8016-75d3a098a58d-kube-api-access-t7hb7\") pod \"ceilometer-0\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " pod="openstack/ceilometer-0" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.947160 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee4992d5-6a42-46d6-8016-75d3a098a58d-run-httpd\") pod \"ceilometer-0\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " pod="openstack/ceilometer-0" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.947178 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee4992d5-6a42-46d6-8016-75d3a098a58d-config-data\") pod \"ceilometer-0\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " pod="openstack/ceilometer-0" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.947197 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee4992d5-6a42-46d6-8016-75d3a098a58d-scripts\") pod \"ceilometer-0\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " pod="openstack/ceilometer-0" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.947263 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee4992d5-6a42-46d6-8016-75d3a098a58d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " pod="openstack/ceilometer-0" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.947343 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee4992d5-6a42-46d6-8016-75d3a098a58d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " pod="openstack/ceilometer-0" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.947371 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee4992d5-6a42-46d6-8016-75d3a098a58d-log-httpd\") pod \"ceilometer-0\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " pod="openstack/ceilometer-0" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.949329 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.949306613 podStartE2EDuration="3.949306613s" podCreationTimestamp="2025-12-03 16:26:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:26:09.882229482 +0000 UTC m=+1348.493929705" watchObservedRunningTime="2025-12-03 16:26:09.949306613 +0000 UTC m=+1348.561006836" Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.977414 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:26:09 crc kubenswrapper[4998]: I1203 16:26:09.990014 4998 scope.go:117] "RemoveContainer" containerID="16a67c27d69bc79ad920a839aa77b4f69a1111da0eb9383f8364f74c2627daa3" Dec 03 16:26:10 crc kubenswrapper[4998]: I1203 16:26:10.018273 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 16:26:10 crc kubenswrapper[4998]: I1203 16:26:10.048719 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee4992d5-6a42-46d6-8016-75d3a098a58d-scripts\") pod \"ceilometer-0\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " pod="openstack/ceilometer-0" Dec 03 16:26:10 crc kubenswrapper[4998]: I1203 16:26:10.048878 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee4992d5-6a42-46d6-8016-75d3a098a58d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " pod="openstack/ceilometer-0" Dec 03 16:26:10 crc kubenswrapper[4998]: I1203 16:26:10.048957 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee4992d5-6a42-46d6-8016-75d3a098a58d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " pod="openstack/ceilometer-0" Dec 03 16:26:10 crc kubenswrapper[4998]: I1203 16:26:10.049014 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee4992d5-6a42-46d6-8016-75d3a098a58d-log-httpd\") pod \"ceilometer-0\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " pod="openstack/ceilometer-0" Dec 03 16:26:10 crc kubenswrapper[4998]: I1203 16:26:10.049052 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7hb7\" (UniqueName: \"kubernetes.io/projected/ee4992d5-6a42-46d6-8016-75d3a098a58d-kube-api-access-t7hb7\") pod \"ceilometer-0\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " pod="openstack/ceilometer-0" Dec 03 16:26:10 crc kubenswrapper[4998]: I1203 16:26:10.049096 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee4992d5-6a42-46d6-8016-75d3a098a58d-run-httpd\") pod \"ceilometer-0\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " pod="openstack/ceilometer-0" Dec 03 16:26:10 crc kubenswrapper[4998]: I1203 16:26:10.049114 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee4992d5-6a42-46d6-8016-75d3a098a58d-config-data\") pod \"ceilometer-0\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " pod="openstack/ceilometer-0" Dec 03 16:26:10 crc kubenswrapper[4998]: I1203 16:26:10.053595 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee4992d5-6a42-46d6-8016-75d3a098a58d-run-httpd\") pod \"ceilometer-0\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " pod="openstack/ceilometer-0" Dec 03 16:26:10 crc kubenswrapper[4998]: I1203 16:26:10.053898 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee4992d5-6a42-46d6-8016-75d3a098a58d-log-httpd\") pod \"ceilometer-0\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " pod="openstack/ceilometer-0" Dec 03 16:26:10 crc kubenswrapper[4998]: I1203 16:26:10.055850 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee4992d5-6a42-46d6-8016-75d3a098a58d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " pod="openstack/ceilometer-0" Dec 03 16:26:10 crc kubenswrapper[4998]: I1203 16:26:10.059063 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee4992d5-6a42-46d6-8016-75d3a098a58d-config-data\") pod \"ceilometer-0\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " pod="openstack/ceilometer-0" Dec 03 16:26:10 crc kubenswrapper[4998]: I1203 16:26:10.061526 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee4992d5-6a42-46d6-8016-75d3a098a58d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " pod="openstack/ceilometer-0" Dec 03 16:26:10 crc kubenswrapper[4998]: I1203 16:26:10.065952 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee4992d5-6a42-46d6-8016-75d3a098a58d-scripts\") pod \"ceilometer-0\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " pod="openstack/ceilometer-0" Dec 03 16:26:10 crc kubenswrapper[4998]: I1203 16:26:10.072471 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7hb7\" (UniqueName: \"kubernetes.io/projected/ee4992d5-6a42-46d6-8016-75d3a098a58d-kube-api-access-t7hb7\") pod \"ceilometer-0\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " pod="openstack/ceilometer-0" Dec 03 16:26:10 crc kubenswrapper[4998]: I1203 16:26:10.268927 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:26:10 crc kubenswrapper[4998]: I1203 16:26:10.669259 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qjtcv" podUID="a78fa691-c7a7-4908-8eb8-b3727b92941d" containerName="registry-server" probeResult="failure" output=< Dec 03 16:26:10 crc kubenswrapper[4998]: timeout: failed to connect service ":50051" within 1s Dec 03 16:26:10 crc kubenswrapper[4998]: > Dec 03 16:26:10 crc kubenswrapper[4998]: I1203 16:26:10.747696 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:26:10 crc kubenswrapper[4998]: W1203 16:26:10.757051 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee4992d5_6a42_46d6_8016_75d3a098a58d.slice/crio-c0deb859b612418cbecaca1e9483921e711115f0307ce5bcb5135564f818c156 WatchSource:0}: Error finding container c0deb859b612418cbecaca1e9483921e711115f0307ce5bcb5135564f818c156: Status 404 returned error can't find the container with id c0deb859b612418cbecaca1e9483921e711115f0307ce5bcb5135564f818c156 Dec 03 16:26:10 crc kubenswrapper[4998]: I1203 16:26:10.857427 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee4992d5-6a42-46d6-8016-75d3a098a58d","Type":"ContainerStarted","Data":"c0deb859b612418cbecaca1e9483921e711115f0307ce5bcb5135564f818c156"} Dec 03 16:26:11 crc kubenswrapper[4998]: I1203 16:26:11.693557 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb" path="/var/lib/kubelet/pods/3a4c9b98-8158-4a8c-a3fe-3e89907a5bbb/volumes" Dec 03 16:26:11 crc kubenswrapper[4998]: I1203 16:26:11.875596 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-decision-engine-0" podUID="a009aeee-c849-4c00-b41b-3cf5e36a75bb" containerName="watcher-decision-engine" containerID="cri-o://c923ec83ef3968e5e7f59971085d59629092a44767625d2597d9b5e9e3e8177e" gracePeriod=30 Dec 03 16:26:11 crc kubenswrapper[4998]: I1203 16:26:11.875899 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee4992d5-6a42-46d6-8016-75d3a098a58d","Type":"ContainerStarted","Data":"16c994d44ef1811789dc2c8d58daf45334192a446b276584615520ebd0396d92"} Dec 03 16:26:11 crc kubenswrapper[4998]: I1203 16:26:11.875927 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee4992d5-6a42-46d6-8016-75d3a098a58d","Type":"ContainerStarted","Data":"fe893386c1daad090003743d48e317a4881b4fea78bf8b40fd6d788116cc7bc7"} Dec 03 16:26:13 crc kubenswrapper[4998]: I1203 16:26:13.902097 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee4992d5-6a42-46d6-8016-75d3a098a58d","Type":"ContainerStarted","Data":"bdd3720d496cc688152df5d9ebda110e3c1a085e1d9bb84835dbdd547388d030"} Dec 03 16:26:14 crc kubenswrapper[4998]: I1203 16:26:14.913141 4998 generic.go:334] "Generic (PLEG): container finished" podID="a009aeee-c849-4c00-b41b-3cf5e36a75bb" containerID="c923ec83ef3968e5e7f59971085d59629092a44767625d2597d9b5e9e3e8177e" exitCode=0 Dec 03 16:26:14 crc kubenswrapper[4998]: I1203 16:26:14.913373 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"a009aeee-c849-4c00-b41b-3cf5e36a75bb","Type":"ContainerDied","Data":"c923ec83ef3968e5e7f59971085d59629092a44767625d2597d9b5e9e3e8177e"} Dec 03 16:26:14 crc kubenswrapper[4998]: I1203 16:26:14.913405 4998 scope.go:117] "RemoveContainer" containerID="b657ed22e9096f43b7242cbed330ab7de4b3e3c8d87141fb5d8555f0560c2457" Dec 03 16:26:15 crc kubenswrapper[4998]: I1203 16:26:15.170908 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 16:26:15 crc kubenswrapper[4998]: I1203 16:26:15.170952 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 16:26:15 crc kubenswrapper[4998]: I1203 16:26:15.206952 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 16:26:15 crc kubenswrapper[4998]: I1203 16:26:15.232323 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 16:26:15 crc kubenswrapper[4998]: I1203 16:26:15.930202 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 16:26:15 crc kubenswrapper[4998]: I1203 16:26:15.930276 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 16:26:17 crc kubenswrapper[4998]: I1203 16:26:17.274716 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 16:26:17 crc kubenswrapper[4998]: I1203 16:26:17.275020 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 16:26:17 crc kubenswrapper[4998]: I1203 16:26:17.331644 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 16:26:17 crc kubenswrapper[4998]: I1203 16:26:17.336239 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 16:26:17 crc kubenswrapper[4998]: I1203 16:26:17.950771 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 16:26:17 crc kubenswrapper[4998]: I1203 16:26:17.950814 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 16:26:18 crc kubenswrapper[4998]: I1203 16:26:18.674036 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 16:26:18 crc kubenswrapper[4998]: I1203 16:26:18.674145 4998 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 16:26:18 crc kubenswrapper[4998]: I1203 16:26:18.696365 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 16:26:19 crc kubenswrapper[4998]: I1203 16:26:19.526204 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 03 16:26:19 crc kubenswrapper[4998]: I1203 16:26:19.680893 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a009aeee-c849-4c00-b41b-3cf5e36a75bb-config-data\") pod \"a009aeee-c849-4c00-b41b-3cf5e36a75bb\" (UID: \"a009aeee-c849-4c00-b41b-3cf5e36a75bb\") " Dec 03 16:26:19 crc kubenswrapper[4998]: I1203 16:26:19.681267 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a009aeee-c849-4c00-b41b-3cf5e36a75bb-logs\") pod \"a009aeee-c849-4c00-b41b-3cf5e36a75bb\" (UID: \"a009aeee-c849-4c00-b41b-3cf5e36a75bb\") " Dec 03 16:26:19 crc kubenswrapper[4998]: I1203 16:26:19.681442 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/a009aeee-c849-4c00-b41b-3cf5e36a75bb-custom-prometheus-ca\") pod \"a009aeee-c849-4c00-b41b-3cf5e36a75bb\" (UID: \"a009aeee-c849-4c00-b41b-3cf5e36a75bb\") " Dec 03 16:26:19 crc kubenswrapper[4998]: I1203 16:26:19.681547 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a009aeee-c849-4c00-b41b-3cf5e36a75bb-combined-ca-bundle\") pod \"a009aeee-c849-4c00-b41b-3cf5e36a75bb\" (UID: \"a009aeee-c849-4c00-b41b-3cf5e36a75bb\") " Dec 03 16:26:19 crc kubenswrapper[4998]: I1203 16:26:19.681624 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcvvt\" (UniqueName: \"kubernetes.io/projected/a009aeee-c849-4c00-b41b-3cf5e36a75bb-kube-api-access-fcvvt\") pod \"a009aeee-c849-4c00-b41b-3cf5e36a75bb\" (UID: \"a009aeee-c849-4c00-b41b-3cf5e36a75bb\") " Dec 03 16:26:19 crc kubenswrapper[4998]: I1203 16:26:19.681906 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a009aeee-c849-4c00-b41b-3cf5e36a75bb-logs" (OuterVolumeSpecName: "logs") pod "a009aeee-c849-4c00-b41b-3cf5e36a75bb" (UID: "a009aeee-c849-4c00-b41b-3cf5e36a75bb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:26:19 crc kubenswrapper[4998]: I1203 16:26:19.682344 4998 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a009aeee-c849-4c00-b41b-3cf5e36a75bb-logs\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:19 crc kubenswrapper[4998]: I1203 16:26:19.697040 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a009aeee-c849-4c00-b41b-3cf5e36a75bb-kube-api-access-fcvvt" (OuterVolumeSpecName: "kube-api-access-fcvvt") pod "a009aeee-c849-4c00-b41b-3cf5e36a75bb" (UID: "a009aeee-c849-4c00-b41b-3cf5e36a75bb"). InnerVolumeSpecName "kube-api-access-fcvvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:26:19 crc kubenswrapper[4998]: I1203 16:26:19.759921 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a009aeee-c849-4c00-b41b-3cf5e36a75bb-config-data" (OuterVolumeSpecName: "config-data") pod "a009aeee-c849-4c00-b41b-3cf5e36a75bb" (UID: "a009aeee-c849-4c00-b41b-3cf5e36a75bb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:19 crc kubenswrapper[4998]: I1203 16:26:19.765867 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a009aeee-c849-4c00-b41b-3cf5e36a75bb-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "a009aeee-c849-4c00-b41b-3cf5e36a75bb" (UID: "a009aeee-c849-4c00-b41b-3cf5e36a75bb"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:19 crc kubenswrapper[4998]: I1203 16:26:19.766206 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a009aeee-c849-4c00-b41b-3cf5e36a75bb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a009aeee-c849-4c00-b41b-3cf5e36a75bb" (UID: "a009aeee-c849-4c00-b41b-3cf5e36a75bb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:19 crc kubenswrapper[4998]: I1203 16:26:19.789454 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a009aeee-c849-4c00-b41b-3cf5e36a75bb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:19 crc kubenswrapper[4998]: I1203 16:26:19.789489 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcvvt\" (UniqueName: \"kubernetes.io/projected/a009aeee-c849-4c00-b41b-3cf5e36a75bb-kube-api-access-fcvvt\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:19 crc kubenswrapper[4998]: I1203 16:26:19.789501 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a009aeee-c849-4c00-b41b-3cf5e36a75bb-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:19 crc kubenswrapper[4998]: I1203 16:26:19.789509 4998 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/a009aeee-c849-4c00-b41b-3cf5e36a75bb-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:19 crc kubenswrapper[4998]: I1203 16:26:19.853454 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qjtcv" Dec 03 16:26:19 crc kubenswrapper[4998]: I1203 16:26:19.941774 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qjtcv" Dec 03 16:26:19 crc kubenswrapper[4998]: I1203 16:26:19.986211 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"a009aeee-c849-4c00-b41b-3cf5e36a75bb","Type":"ContainerDied","Data":"510d649767f5e1e8480ee19550b358b84232ec13dd0041d57c75b1d5574ef8b3"} Dec 03 16:26:19 crc kubenswrapper[4998]: I1203 16:26:19.986232 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 03 16:26:19 crc kubenswrapper[4998]: I1203 16:26:19.986258 4998 scope.go:117] "RemoveContainer" containerID="c923ec83ef3968e5e7f59971085d59629092a44767625d2597d9b5e9e3e8177e" Dec 03 16:26:19 crc kubenswrapper[4998]: I1203 16:26:19.993409 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6zkx5" event={"ID":"3ff4ba4e-9b43-41fd-902b-977b9cc3aba9","Type":"ContainerStarted","Data":"c4fe18fac933af612bc197d41649650f29ab6e8496268bd2342d1c84935c7234"} Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.072399 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.079992 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.089815 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 16:26:20 crc kubenswrapper[4998]: E1203 16:26:20.090259 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a009aeee-c849-4c00-b41b-3cf5e36a75bb" containerName="watcher-decision-engine" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.090274 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="a009aeee-c849-4c00-b41b-3cf5e36a75bb" containerName="watcher-decision-engine" Dec 03 16:26:20 crc kubenswrapper[4998]: E1203 16:26:20.090286 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a009aeee-c849-4c00-b41b-3cf5e36a75bb" containerName="watcher-decision-engine" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.090294 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="a009aeee-c849-4c00-b41b-3cf5e36a75bb" containerName="watcher-decision-engine" Dec 03 16:26:20 crc kubenswrapper[4998]: E1203 16:26:20.090312 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a009aeee-c849-4c00-b41b-3cf5e36a75bb" containerName="watcher-decision-engine" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.090322 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="a009aeee-c849-4c00-b41b-3cf5e36a75bb" containerName="watcher-decision-engine" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.090516 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="a009aeee-c849-4c00-b41b-3cf5e36a75bb" containerName="watcher-decision-engine" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.090544 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="a009aeee-c849-4c00-b41b-3cf5e36a75bb" containerName="watcher-decision-engine" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.090558 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="a009aeee-c849-4c00-b41b-3cf5e36a75bb" containerName="watcher-decision-engine" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.091396 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.095035 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-6zkx5" podStartSLOduration=2.3868070599999998 podStartE2EDuration="13.095017802s" podCreationTimestamp="2025-12-03 16:26:07 +0000 UTC" firstStartedPulling="2025-12-03 16:26:08.841809213 +0000 UTC m=+1347.453509436" lastFinishedPulling="2025-12-03 16:26:19.550019955 +0000 UTC m=+1358.161720178" observedRunningTime="2025-12-03 16:26:20.052011416 +0000 UTC m=+1358.663711639" watchObservedRunningTime="2025-12-03 16:26:20.095017802 +0000 UTC m=+1358.706718025" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.105639 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.111435 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.145254 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qjtcv"] Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.198955 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aaa74904-2894-4cd4-9c3d-6080c69664db-logs\") pod \"watcher-decision-engine-0\" (UID: \"aaa74904-2894-4cd4-9c3d-6080c69664db\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.199048 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaa74904-2894-4cd4-9c3d-6080c69664db-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"aaa74904-2894-4cd4-9c3d-6080c69664db\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.199108 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaa74904-2894-4cd4-9c3d-6080c69664db-config-data\") pod \"watcher-decision-engine-0\" (UID: \"aaa74904-2894-4cd4-9c3d-6080c69664db\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.199211 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/aaa74904-2894-4cd4-9c3d-6080c69664db-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"aaa74904-2894-4cd4-9c3d-6080c69664db\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.199250 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8lb8\" (UniqueName: \"kubernetes.io/projected/aaa74904-2894-4cd4-9c3d-6080c69664db-kube-api-access-p8lb8\") pod \"watcher-decision-engine-0\" (UID: \"aaa74904-2894-4cd4-9c3d-6080c69664db\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.301122 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaa74904-2894-4cd4-9c3d-6080c69664db-config-data\") pod \"watcher-decision-engine-0\" (UID: \"aaa74904-2894-4cd4-9c3d-6080c69664db\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.301523 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/aaa74904-2894-4cd4-9c3d-6080c69664db-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"aaa74904-2894-4cd4-9c3d-6080c69664db\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.301647 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8lb8\" (UniqueName: \"kubernetes.io/projected/aaa74904-2894-4cd4-9c3d-6080c69664db-kube-api-access-p8lb8\") pod \"watcher-decision-engine-0\" (UID: \"aaa74904-2894-4cd4-9c3d-6080c69664db\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.301786 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aaa74904-2894-4cd4-9c3d-6080c69664db-logs\") pod \"watcher-decision-engine-0\" (UID: \"aaa74904-2894-4cd4-9c3d-6080c69664db\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.301965 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaa74904-2894-4cd4-9c3d-6080c69664db-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"aaa74904-2894-4cd4-9c3d-6080c69664db\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.302256 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aaa74904-2894-4cd4-9c3d-6080c69664db-logs\") pod \"watcher-decision-engine-0\" (UID: \"aaa74904-2894-4cd4-9c3d-6080c69664db\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.307293 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaa74904-2894-4cd4-9c3d-6080c69664db-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"aaa74904-2894-4cd4-9c3d-6080c69664db\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.311471 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaa74904-2894-4cd4-9c3d-6080c69664db-config-data\") pod \"watcher-decision-engine-0\" (UID: \"aaa74904-2894-4cd4-9c3d-6080c69664db\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.313364 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/aaa74904-2894-4cd4-9c3d-6080c69664db-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"aaa74904-2894-4cd4-9c3d-6080c69664db\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.321250 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8lb8\" (UniqueName: \"kubernetes.io/projected/aaa74904-2894-4cd4-9c3d-6080c69664db-kube-api-access-p8lb8\") pod \"watcher-decision-engine-0\" (UID: \"aaa74904-2894-4cd4-9c3d-6080c69664db\") " pod="openstack/watcher-decision-engine-0" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.419923 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.420581 4998 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.427720 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.610286 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 16:26:20 crc kubenswrapper[4998]: I1203 16:26:20.988975 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 03 16:26:21 crc kubenswrapper[4998]: I1203 16:26:21.009077 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee4992d5-6a42-46d6-8016-75d3a098a58d","Type":"ContainerStarted","Data":"f4f57ee9ef9fd18987c95cea5e3c4c430417635ff89c89e2d3d100e9375d995c"} Dec 03 16:26:21 crc kubenswrapper[4998]: I1203 16:26:21.009564 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 16:26:21 crc kubenswrapper[4998]: W1203 16:26:21.010131 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaaa74904_2894_4cd4_9c3d_6080c69664db.slice/crio-2760d8228dbfd2204ddf4b08c795568690877554dde1614f694d5f64f9c3a1a3 WatchSource:0}: Error finding container 2760d8228dbfd2204ddf4b08c795568690877554dde1614f694d5f64f9c3a1a3: Status 404 returned error can't find the container with id 2760d8228dbfd2204ddf4b08c795568690877554dde1614f694d5f64f9c3a1a3 Dec 03 16:26:21 crc kubenswrapper[4998]: I1203 16:26:21.011589 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qjtcv" podUID="a78fa691-c7a7-4908-8eb8-b3727b92941d" containerName="registry-server" containerID="cri-o://d90119e30812730816317a3ca8fc16f0f405cd659a5eaee1babf027e432b911d" gracePeriod=2 Dec 03 16:26:21 crc kubenswrapper[4998]: I1203 16:26:21.039660 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.2518781199999998 podStartE2EDuration="12.039640556s" podCreationTimestamp="2025-12-03 16:26:09 +0000 UTC" firstStartedPulling="2025-12-03 16:26:10.760142078 +0000 UTC m=+1349.371842301" lastFinishedPulling="2025-12-03 16:26:19.547904514 +0000 UTC m=+1358.159604737" observedRunningTime="2025-12-03 16:26:21.029239944 +0000 UTC m=+1359.640940197" watchObservedRunningTime="2025-12-03 16:26:21.039640556 +0000 UTC m=+1359.651340779" Dec 03 16:26:21 crc kubenswrapper[4998]: I1203 16:26:21.427910 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qjtcv" Dec 03 16:26:21 crc kubenswrapper[4998]: I1203 16:26:21.536615 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wf84\" (UniqueName: \"kubernetes.io/projected/a78fa691-c7a7-4908-8eb8-b3727b92941d-kube-api-access-9wf84\") pod \"a78fa691-c7a7-4908-8eb8-b3727b92941d\" (UID: \"a78fa691-c7a7-4908-8eb8-b3727b92941d\") " Dec 03 16:26:21 crc kubenswrapper[4998]: I1203 16:26:21.536919 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a78fa691-c7a7-4908-8eb8-b3727b92941d-utilities\") pod \"a78fa691-c7a7-4908-8eb8-b3727b92941d\" (UID: \"a78fa691-c7a7-4908-8eb8-b3727b92941d\") " Dec 03 16:26:21 crc kubenswrapper[4998]: I1203 16:26:21.537035 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a78fa691-c7a7-4908-8eb8-b3727b92941d-catalog-content\") pod \"a78fa691-c7a7-4908-8eb8-b3727b92941d\" (UID: \"a78fa691-c7a7-4908-8eb8-b3727b92941d\") " Dec 03 16:26:21 crc kubenswrapper[4998]: I1203 16:26:21.538198 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a78fa691-c7a7-4908-8eb8-b3727b92941d-utilities" (OuterVolumeSpecName: "utilities") pod "a78fa691-c7a7-4908-8eb8-b3727b92941d" (UID: "a78fa691-c7a7-4908-8eb8-b3727b92941d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:26:21 crc kubenswrapper[4998]: I1203 16:26:21.540950 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a78fa691-c7a7-4908-8eb8-b3727b92941d-kube-api-access-9wf84" (OuterVolumeSpecName: "kube-api-access-9wf84") pod "a78fa691-c7a7-4908-8eb8-b3727b92941d" (UID: "a78fa691-c7a7-4908-8eb8-b3727b92941d"). InnerVolumeSpecName "kube-api-access-9wf84". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:26:21 crc kubenswrapper[4998]: I1203 16:26:21.632459 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a78fa691-c7a7-4908-8eb8-b3727b92941d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a78fa691-c7a7-4908-8eb8-b3727b92941d" (UID: "a78fa691-c7a7-4908-8eb8-b3727b92941d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:26:21 crc kubenswrapper[4998]: I1203 16:26:21.639420 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wf84\" (UniqueName: \"kubernetes.io/projected/a78fa691-c7a7-4908-8eb8-b3727b92941d-kube-api-access-9wf84\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:21 crc kubenswrapper[4998]: I1203 16:26:21.639460 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a78fa691-c7a7-4908-8eb8-b3727b92941d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:21 crc kubenswrapper[4998]: I1203 16:26:21.639475 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a78fa691-c7a7-4908-8eb8-b3727b92941d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:21 crc kubenswrapper[4998]: I1203 16:26:21.691428 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a009aeee-c849-4c00-b41b-3cf5e36a75bb" path="/var/lib/kubelet/pods/a009aeee-c849-4c00-b41b-3cf5e36a75bb/volumes" Dec 03 16:26:22 crc kubenswrapper[4998]: I1203 16:26:22.022058 4998 generic.go:334] "Generic (PLEG): container finished" podID="a78fa691-c7a7-4908-8eb8-b3727b92941d" containerID="d90119e30812730816317a3ca8fc16f0f405cd659a5eaee1babf027e432b911d" exitCode=0 Dec 03 16:26:22 crc kubenswrapper[4998]: I1203 16:26:22.022119 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qjtcv" Dec 03 16:26:22 crc kubenswrapper[4998]: I1203 16:26:22.022127 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qjtcv" event={"ID":"a78fa691-c7a7-4908-8eb8-b3727b92941d","Type":"ContainerDied","Data":"d90119e30812730816317a3ca8fc16f0f405cd659a5eaee1babf027e432b911d"} Dec 03 16:26:22 crc kubenswrapper[4998]: I1203 16:26:22.022156 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qjtcv" event={"ID":"a78fa691-c7a7-4908-8eb8-b3727b92941d","Type":"ContainerDied","Data":"edc42995ad9c0e08eeb54a10a5dd7ecaec8751942eea70f64e501ccd9c4730c1"} Dec 03 16:26:22 crc kubenswrapper[4998]: I1203 16:26:22.022179 4998 scope.go:117] "RemoveContainer" containerID="d90119e30812730816317a3ca8fc16f0f405cd659a5eaee1babf027e432b911d" Dec 03 16:26:22 crc kubenswrapper[4998]: I1203 16:26:22.026315 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"aaa74904-2894-4cd4-9c3d-6080c69664db","Type":"ContainerStarted","Data":"7662511ef9c4475fea41ee6a62e56d237c0d63441dd2cc93f562f0e8a5e3a4d6"} Dec 03 16:26:22 crc kubenswrapper[4998]: I1203 16:26:22.026349 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"aaa74904-2894-4cd4-9c3d-6080c69664db","Type":"ContainerStarted","Data":"2760d8228dbfd2204ddf4b08c795568690877554dde1614f694d5f64f9c3a1a3"} Dec 03 16:26:22 crc kubenswrapper[4998]: I1203 16:26:22.051262 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=2.051248556 podStartE2EDuration="2.051248556s" podCreationTimestamp="2025-12-03 16:26:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:26:22.051047793 +0000 UTC m=+1360.662748016" watchObservedRunningTime="2025-12-03 16:26:22.051248556 +0000 UTC m=+1360.662948779" Dec 03 16:26:22 crc kubenswrapper[4998]: I1203 16:26:22.055047 4998 scope.go:117] "RemoveContainer" containerID="f5fda72cc17b1098cc27f616651a1083ad5dc310101edb27f5380804017543e6" Dec 03 16:26:22 crc kubenswrapper[4998]: I1203 16:26:22.089935 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qjtcv"] Dec 03 16:26:22 crc kubenswrapper[4998]: I1203 16:26:22.093297 4998 scope.go:117] "RemoveContainer" containerID="863de87c4eb155ccc524703ea1a7fb18595e6f92612c8969ca41517db5289aa5" Dec 03 16:26:22 crc kubenswrapper[4998]: I1203 16:26:22.144990 4998 scope.go:117] "RemoveContainer" containerID="d90119e30812730816317a3ca8fc16f0f405cd659a5eaee1babf027e432b911d" Dec 03 16:26:22 crc kubenswrapper[4998]: E1203 16:26:22.145529 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d90119e30812730816317a3ca8fc16f0f405cd659a5eaee1babf027e432b911d\": container with ID starting with d90119e30812730816317a3ca8fc16f0f405cd659a5eaee1babf027e432b911d not found: ID does not exist" containerID="d90119e30812730816317a3ca8fc16f0f405cd659a5eaee1babf027e432b911d" Dec 03 16:26:22 crc kubenswrapper[4998]: I1203 16:26:22.145559 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d90119e30812730816317a3ca8fc16f0f405cd659a5eaee1babf027e432b911d"} err="failed to get container status \"d90119e30812730816317a3ca8fc16f0f405cd659a5eaee1babf027e432b911d\": rpc error: code = NotFound desc = could not find container \"d90119e30812730816317a3ca8fc16f0f405cd659a5eaee1babf027e432b911d\": container with ID starting with d90119e30812730816317a3ca8fc16f0f405cd659a5eaee1babf027e432b911d not found: ID does not exist" Dec 03 16:26:22 crc kubenswrapper[4998]: I1203 16:26:22.145591 4998 scope.go:117] "RemoveContainer" containerID="f5fda72cc17b1098cc27f616651a1083ad5dc310101edb27f5380804017543e6" Dec 03 16:26:22 crc kubenswrapper[4998]: E1203 16:26:22.147109 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5fda72cc17b1098cc27f616651a1083ad5dc310101edb27f5380804017543e6\": container with ID starting with f5fda72cc17b1098cc27f616651a1083ad5dc310101edb27f5380804017543e6 not found: ID does not exist" containerID="f5fda72cc17b1098cc27f616651a1083ad5dc310101edb27f5380804017543e6" Dec 03 16:26:22 crc kubenswrapper[4998]: I1203 16:26:22.147148 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5fda72cc17b1098cc27f616651a1083ad5dc310101edb27f5380804017543e6"} err="failed to get container status \"f5fda72cc17b1098cc27f616651a1083ad5dc310101edb27f5380804017543e6\": rpc error: code = NotFound desc = could not find container \"f5fda72cc17b1098cc27f616651a1083ad5dc310101edb27f5380804017543e6\": container with ID starting with f5fda72cc17b1098cc27f616651a1083ad5dc310101edb27f5380804017543e6 not found: ID does not exist" Dec 03 16:26:22 crc kubenswrapper[4998]: I1203 16:26:22.147165 4998 scope.go:117] "RemoveContainer" containerID="863de87c4eb155ccc524703ea1a7fb18595e6f92612c8969ca41517db5289aa5" Dec 03 16:26:22 crc kubenswrapper[4998]: E1203 16:26:22.151390 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"863de87c4eb155ccc524703ea1a7fb18595e6f92612c8969ca41517db5289aa5\": container with ID starting with 863de87c4eb155ccc524703ea1a7fb18595e6f92612c8969ca41517db5289aa5 not found: ID does not exist" containerID="863de87c4eb155ccc524703ea1a7fb18595e6f92612c8969ca41517db5289aa5" Dec 03 16:26:22 crc kubenswrapper[4998]: I1203 16:26:22.151431 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"863de87c4eb155ccc524703ea1a7fb18595e6f92612c8969ca41517db5289aa5"} err="failed to get container status \"863de87c4eb155ccc524703ea1a7fb18595e6f92612c8969ca41517db5289aa5\": rpc error: code = NotFound desc = could not find container \"863de87c4eb155ccc524703ea1a7fb18595e6f92612c8969ca41517db5289aa5\": container with ID starting with 863de87c4eb155ccc524703ea1a7fb18595e6f92612c8969ca41517db5289aa5 not found: ID does not exist" Dec 03 16:26:22 crc kubenswrapper[4998]: I1203 16:26:22.181660 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qjtcv"] Dec 03 16:26:23 crc kubenswrapper[4998]: I1203 16:26:23.690945 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a78fa691-c7a7-4908-8eb8-b3727b92941d" path="/var/lib/kubelet/pods/a78fa691-c7a7-4908-8eb8-b3727b92941d/volumes" Dec 03 16:26:24 crc kubenswrapper[4998]: I1203 16:26:24.484401 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:26:24 crc kubenswrapper[4998]: I1203 16:26:24.484970 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee4992d5-6a42-46d6-8016-75d3a098a58d" containerName="ceilometer-central-agent" containerID="cri-o://fe893386c1daad090003743d48e317a4881b4fea78bf8b40fd6d788116cc7bc7" gracePeriod=30 Dec 03 16:26:24 crc kubenswrapper[4998]: I1203 16:26:24.485384 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee4992d5-6a42-46d6-8016-75d3a098a58d" containerName="proxy-httpd" containerID="cri-o://f4f57ee9ef9fd18987c95cea5e3c4c430417635ff89c89e2d3d100e9375d995c" gracePeriod=30 Dec 03 16:26:24 crc kubenswrapper[4998]: I1203 16:26:24.485445 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee4992d5-6a42-46d6-8016-75d3a098a58d" containerName="sg-core" containerID="cri-o://bdd3720d496cc688152df5d9ebda110e3c1a085e1d9bb84835dbdd547388d030" gracePeriod=30 Dec 03 16:26:24 crc kubenswrapper[4998]: I1203 16:26:24.485480 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ee4992d5-6a42-46d6-8016-75d3a098a58d" containerName="ceilometer-notification-agent" containerID="cri-o://16c994d44ef1811789dc2c8d58daf45334192a446b276584615520ebd0396d92" gracePeriod=30 Dec 03 16:26:25 crc kubenswrapper[4998]: I1203 16:26:25.104178 4998 generic.go:334] "Generic (PLEG): container finished" podID="ee4992d5-6a42-46d6-8016-75d3a098a58d" containerID="f4f57ee9ef9fd18987c95cea5e3c4c430417635ff89c89e2d3d100e9375d995c" exitCode=0 Dec 03 16:26:25 crc kubenswrapper[4998]: I1203 16:26:25.104223 4998 generic.go:334] "Generic (PLEG): container finished" podID="ee4992d5-6a42-46d6-8016-75d3a098a58d" containerID="bdd3720d496cc688152df5d9ebda110e3c1a085e1d9bb84835dbdd547388d030" exitCode=2 Dec 03 16:26:25 crc kubenswrapper[4998]: I1203 16:26:25.104237 4998 generic.go:334] "Generic (PLEG): container finished" podID="ee4992d5-6a42-46d6-8016-75d3a098a58d" containerID="fe893386c1daad090003743d48e317a4881b4fea78bf8b40fd6d788116cc7bc7" exitCode=0 Dec 03 16:26:25 crc kubenswrapper[4998]: I1203 16:26:25.104263 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee4992d5-6a42-46d6-8016-75d3a098a58d","Type":"ContainerDied","Data":"f4f57ee9ef9fd18987c95cea5e3c4c430417635ff89c89e2d3d100e9375d995c"} Dec 03 16:26:25 crc kubenswrapper[4998]: I1203 16:26:25.104298 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee4992d5-6a42-46d6-8016-75d3a098a58d","Type":"ContainerDied","Data":"bdd3720d496cc688152df5d9ebda110e3c1a085e1d9bb84835dbdd547388d030"} Dec 03 16:26:25 crc kubenswrapper[4998]: I1203 16:26:25.104313 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee4992d5-6a42-46d6-8016-75d3a098a58d","Type":"ContainerDied","Data":"fe893386c1daad090003743d48e317a4881b4fea78bf8b40fd6d788116cc7bc7"} Dec 03 16:26:29 crc kubenswrapper[4998]: I1203 16:26:29.157727 4998 generic.go:334] "Generic (PLEG): container finished" podID="ee4992d5-6a42-46d6-8016-75d3a098a58d" containerID="16c994d44ef1811789dc2c8d58daf45334192a446b276584615520ebd0396d92" exitCode=0 Dec 03 16:26:29 crc kubenswrapper[4998]: I1203 16:26:29.157802 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee4992d5-6a42-46d6-8016-75d3a098a58d","Type":"ContainerDied","Data":"16c994d44ef1811789dc2c8d58daf45334192a446b276584615520ebd0396d92"} Dec 03 16:26:29 crc kubenswrapper[4998]: I1203 16:26:29.519323 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:26:29 crc kubenswrapper[4998]: I1203 16:26:29.697545 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7hb7\" (UniqueName: \"kubernetes.io/projected/ee4992d5-6a42-46d6-8016-75d3a098a58d-kube-api-access-t7hb7\") pod \"ee4992d5-6a42-46d6-8016-75d3a098a58d\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " Dec 03 16:26:29 crc kubenswrapper[4998]: I1203 16:26:29.697683 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee4992d5-6a42-46d6-8016-75d3a098a58d-run-httpd\") pod \"ee4992d5-6a42-46d6-8016-75d3a098a58d\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " Dec 03 16:26:29 crc kubenswrapper[4998]: I1203 16:26:29.698072 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee4992d5-6a42-46d6-8016-75d3a098a58d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ee4992d5-6a42-46d6-8016-75d3a098a58d" (UID: "ee4992d5-6a42-46d6-8016-75d3a098a58d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:26:29 crc kubenswrapper[4998]: I1203 16:26:29.698140 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee4992d5-6a42-46d6-8016-75d3a098a58d-sg-core-conf-yaml\") pod \"ee4992d5-6a42-46d6-8016-75d3a098a58d\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " Dec 03 16:26:29 crc kubenswrapper[4998]: I1203 16:26:29.698197 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee4992d5-6a42-46d6-8016-75d3a098a58d-log-httpd\") pod \"ee4992d5-6a42-46d6-8016-75d3a098a58d\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " Dec 03 16:26:29 crc kubenswrapper[4998]: I1203 16:26:29.698240 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee4992d5-6a42-46d6-8016-75d3a098a58d-combined-ca-bundle\") pod \"ee4992d5-6a42-46d6-8016-75d3a098a58d\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " Dec 03 16:26:29 crc kubenswrapper[4998]: I1203 16:26:29.698271 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee4992d5-6a42-46d6-8016-75d3a098a58d-config-data\") pod \"ee4992d5-6a42-46d6-8016-75d3a098a58d\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " Dec 03 16:26:29 crc kubenswrapper[4998]: I1203 16:26:29.698348 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee4992d5-6a42-46d6-8016-75d3a098a58d-scripts\") pod \"ee4992d5-6a42-46d6-8016-75d3a098a58d\" (UID: \"ee4992d5-6a42-46d6-8016-75d3a098a58d\") " Dec 03 16:26:29 crc kubenswrapper[4998]: I1203 16:26:29.698636 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee4992d5-6a42-46d6-8016-75d3a098a58d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ee4992d5-6a42-46d6-8016-75d3a098a58d" (UID: "ee4992d5-6a42-46d6-8016-75d3a098a58d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:26:29 crc kubenswrapper[4998]: I1203 16:26:29.698875 4998 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee4992d5-6a42-46d6-8016-75d3a098a58d-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:29 crc kubenswrapper[4998]: I1203 16:26:29.698895 4998 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee4992d5-6a42-46d6-8016-75d3a098a58d-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:29 crc kubenswrapper[4998]: I1203 16:26:29.704263 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee4992d5-6a42-46d6-8016-75d3a098a58d-scripts" (OuterVolumeSpecName: "scripts") pod "ee4992d5-6a42-46d6-8016-75d3a098a58d" (UID: "ee4992d5-6a42-46d6-8016-75d3a098a58d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:29 crc kubenswrapper[4998]: I1203 16:26:29.704905 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee4992d5-6a42-46d6-8016-75d3a098a58d-kube-api-access-t7hb7" (OuterVolumeSpecName: "kube-api-access-t7hb7") pod "ee4992d5-6a42-46d6-8016-75d3a098a58d" (UID: "ee4992d5-6a42-46d6-8016-75d3a098a58d"). InnerVolumeSpecName "kube-api-access-t7hb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:26:29 crc kubenswrapper[4998]: I1203 16:26:29.730930 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee4992d5-6a42-46d6-8016-75d3a098a58d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ee4992d5-6a42-46d6-8016-75d3a098a58d" (UID: "ee4992d5-6a42-46d6-8016-75d3a098a58d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:29 crc kubenswrapper[4998]: I1203 16:26:29.772276 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee4992d5-6a42-46d6-8016-75d3a098a58d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee4992d5-6a42-46d6-8016-75d3a098a58d" (UID: "ee4992d5-6a42-46d6-8016-75d3a098a58d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:29 crc kubenswrapper[4998]: I1203 16:26:29.798821 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee4992d5-6a42-46d6-8016-75d3a098a58d-config-data" (OuterVolumeSpecName: "config-data") pod "ee4992d5-6a42-46d6-8016-75d3a098a58d" (UID: "ee4992d5-6a42-46d6-8016-75d3a098a58d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:29 crc kubenswrapper[4998]: I1203 16:26:29.800612 4998 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ee4992d5-6a42-46d6-8016-75d3a098a58d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:29 crc kubenswrapper[4998]: I1203 16:26:29.800637 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee4992d5-6a42-46d6-8016-75d3a098a58d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:29 crc kubenswrapper[4998]: I1203 16:26:29.800648 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee4992d5-6a42-46d6-8016-75d3a098a58d-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:29 crc kubenswrapper[4998]: I1203 16:26:29.800657 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee4992d5-6a42-46d6-8016-75d3a098a58d-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:29 crc kubenswrapper[4998]: I1203 16:26:29.800666 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7hb7\" (UniqueName: \"kubernetes.io/projected/ee4992d5-6a42-46d6-8016-75d3a098a58d-kube-api-access-t7hb7\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.174972 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ee4992d5-6a42-46d6-8016-75d3a098a58d","Type":"ContainerDied","Data":"c0deb859b612418cbecaca1e9483921e711115f0307ce5bcb5135564f818c156"} Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.175044 4998 scope.go:117] "RemoveContainer" containerID="f4f57ee9ef9fd18987c95cea5e3c4c430417635ff89c89e2d3d100e9375d995c" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.175083 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.197667 4998 scope.go:117] "RemoveContainer" containerID="bdd3720d496cc688152df5d9ebda110e3c1a085e1d9bb84835dbdd547388d030" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.212127 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.219385 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.223533 4998 scope.go:117] "RemoveContainer" containerID="16c994d44ef1811789dc2c8d58daf45334192a446b276584615520ebd0396d92" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.241446 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:26:30 crc kubenswrapper[4998]: E1203 16:26:30.241916 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee4992d5-6a42-46d6-8016-75d3a098a58d" containerName="ceilometer-notification-agent" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.241935 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee4992d5-6a42-46d6-8016-75d3a098a58d" containerName="ceilometer-notification-agent" Dec 03 16:26:30 crc kubenswrapper[4998]: E1203 16:26:30.241950 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee4992d5-6a42-46d6-8016-75d3a098a58d" containerName="sg-core" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.241957 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee4992d5-6a42-46d6-8016-75d3a098a58d" containerName="sg-core" Dec 03 16:26:30 crc kubenswrapper[4998]: E1203 16:26:30.241974 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a78fa691-c7a7-4908-8eb8-b3727b92941d" containerName="extract-utilities" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.241982 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="a78fa691-c7a7-4908-8eb8-b3727b92941d" containerName="extract-utilities" Dec 03 16:26:30 crc kubenswrapper[4998]: E1203 16:26:30.241990 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee4992d5-6a42-46d6-8016-75d3a098a58d" containerName="ceilometer-central-agent" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.241996 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee4992d5-6a42-46d6-8016-75d3a098a58d" containerName="ceilometer-central-agent" Dec 03 16:26:30 crc kubenswrapper[4998]: E1203 16:26:30.242018 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a009aeee-c849-4c00-b41b-3cf5e36a75bb" containerName="watcher-decision-engine" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.242024 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="a009aeee-c849-4c00-b41b-3cf5e36a75bb" containerName="watcher-decision-engine" Dec 03 16:26:30 crc kubenswrapper[4998]: E1203 16:26:30.242036 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee4992d5-6a42-46d6-8016-75d3a098a58d" containerName="proxy-httpd" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.242044 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee4992d5-6a42-46d6-8016-75d3a098a58d" containerName="proxy-httpd" Dec 03 16:26:30 crc kubenswrapper[4998]: E1203 16:26:30.242056 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a78fa691-c7a7-4908-8eb8-b3727b92941d" containerName="extract-content" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.242062 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="a78fa691-c7a7-4908-8eb8-b3727b92941d" containerName="extract-content" Dec 03 16:26:30 crc kubenswrapper[4998]: E1203 16:26:30.242073 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a78fa691-c7a7-4908-8eb8-b3727b92941d" containerName="registry-server" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.242078 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="a78fa691-c7a7-4908-8eb8-b3727b92941d" containerName="registry-server" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.242271 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee4992d5-6a42-46d6-8016-75d3a098a58d" containerName="ceilometer-central-agent" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.242291 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee4992d5-6a42-46d6-8016-75d3a098a58d" containerName="ceilometer-notification-agent" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.242309 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee4992d5-6a42-46d6-8016-75d3a098a58d" containerName="sg-core" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.242322 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="a78fa691-c7a7-4908-8eb8-b3727b92941d" containerName="registry-server" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.242339 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee4992d5-6a42-46d6-8016-75d3a098a58d" containerName="proxy-httpd" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.242354 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="a009aeee-c849-4c00-b41b-3cf5e36a75bb" containerName="watcher-decision-engine" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.244631 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.250630 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.250826 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.259913 4998 scope.go:117] "RemoveContainer" containerID="fe893386c1daad090003743d48e317a4881b4fea78bf8b40fd6d788116cc7bc7" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.265919 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.412479 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17475de4-e997-47c0-83be-9a6f9e5f9857-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " pod="openstack/ceilometer-0" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.412535 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17475de4-e997-47c0-83be-9a6f9e5f9857-run-httpd\") pod \"ceilometer-0\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " pod="openstack/ceilometer-0" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.412576 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17475de4-e997-47c0-83be-9a6f9e5f9857-scripts\") pod \"ceilometer-0\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " pod="openstack/ceilometer-0" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.412600 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t68lc\" (UniqueName: \"kubernetes.io/projected/17475de4-e997-47c0-83be-9a6f9e5f9857-kube-api-access-t68lc\") pod \"ceilometer-0\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " pod="openstack/ceilometer-0" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.412657 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/17475de4-e997-47c0-83be-9a6f9e5f9857-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " pod="openstack/ceilometer-0" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.412945 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17475de4-e997-47c0-83be-9a6f9e5f9857-config-data\") pod \"ceilometer-0\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " pod="openstack/ceilometer-0" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.413083 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17475de4-e997-47c0-83be-9a6f9e5f9857-log-httpd\") pod \"ceilometer-0\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " pod="openstack/ceilometer-0" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.428398 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.457315 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.516052 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17475de4-e997-47c0-83be-9a6f9e5f9857-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " pod="openstack/ceilometer-0" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.516423 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17475de4-e997-47c0-83be-9a6f9e5f9857-run-httpd\") pod \"ceilometer-0\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " pod="openstack/ceilometer-0" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.516605 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17475de4-e997-47c0-83be-9a6f9e5f9857-scripts\") pod \"ceilometer-0\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " pod="openstack/ceilometer-0" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.516845 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t68lc\" (UniqueName: \"kubernetes.io/projected/17475de4-e997-47c0-83be-9a6f9e5f9857-kube-api-access-t68lc\") pod \"ceilometer-0\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " pod="openstack/ceilometer-0" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.517429 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/17475de4-e997-47c0-83be-9a6f9e5f9857-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " pod="openstack/ceilometer-0" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.517621 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17475de4-e997-47c0-83be-9a6f9e5f9857-config-data\") pod \"ceilometer-0\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " pod="openstack/ceilometer-0" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.517745 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17475de4-e997-47c0-83be-9a6f9e5f9857-log-httpd\") pod \"ceilometer-0\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " pod="openstack/ceilometer-0" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.517859 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17475de4-e997-47c0-83be-9a6f9e5f9857-run-httpd\") pod \"ceilometer-0\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " pod="openstack/ceilometer-0" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.518231 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17475de4-e997-47c0-83be-9a6f9e5f9857-log-httpd\") pod \"ceilometer-0\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " pod="openstack/ceilometer-0" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.522581 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/17475de4-e997-47c0-83be-9a6f9e5f9857-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " pod="openstack/ceilometer-0" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.524892 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17475de4-e997-47c0-83be-9a6f9e5f9857-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " pod="openstack/ceilometer-0" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.526089 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17475de4-e997-47c0-83be-9a6f9e5f9857-config-data\") pod \"ceilometer-0\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " pod="openstack/ceilometer-0" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.530517 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17475de4-e997-47c0-83be-9a6f9e5f9857-scripts\") pod \"ceilometer-0\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " pod="openstack/ceilometer-0" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.546573 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t68lc\" (UniqueName: \"kubernetes.io/projected/17475de4-e997-47c0-83be-9a6f9e5f9857-kube-api-access-t68lc\") pod \"ceilometer-0\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " pod="openstack/ceilometer-0" Dec 03 16:26:30 crc kubenswrapper[4998]: I1203 16:26:30.576437 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:26:31 crc kubenswrapper[4998]: I1203 16:26:31.088570 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:26:31 crc kubenswrapper[4998]: W1203 16:26:31.101446 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17475de4_e997_47c0_83be_9a6f9e5f9857.slice/crio-ac3426c408f3e150921fdf71b2b1ed57b048a0ed7f8c4ff804a9d769378b274e WatchSource:0}: Error finding container ac3426c408f3e150921fdf71b2b1ed57b048a0ed7f8c4ff804a9d769378b274e: Status 404 returned error can't find the container with id ac3426c408f3e150921fdf71b2b1ed57b048a0ed7f8c4ff804a9d769378b274e Dec 03 16:26:31 crc kubenswrapper[4998]: I1203 16:26:31.186203 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17475de4-e997-47c0-83be-9a6f9e5f9857","Type":"ContainerStarted","Data":"ac3426c408f3e150921fdf71b2b1ed57b048a0ed7f8c4ff804a9d769378b274e"} Dec 03 16:26:31 crc kubenswrapper[4998]: I1203 16:26:31.187726 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Dec 03 16:26:31 crc kubenswrapper[4998]: I1203 16:26:31.242400 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Dec 03 16:26:31 crc kubenswrapper[4998]: I1203 16:26:31.696083 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee4992d5-6a42-46d6-8016-75d3a098a58d" path="/var/lib/kubelet/pods/ee4992d5-6a42-46d6-8016-75d3a098a58d/volumes" Dec 03 16:26:33 crc kubenswrapper[4998]: I1203 16:26:33.207895 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17475de4-e997-47c0-83be-9a6f9e5f9857","Type":"ContainerStarted","Data":"9d370332bd21a0f3c8106952dc5a5162b551ed091e5b56a0624befd147c8cf34"} Dec 03 16:26:33 crc kubenswrapper[4998]: I1203 16:26:33.208302 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17475de4-e997-47c0-83be-9a6f9e5f9857","Type":"ContainerStarted","Data":"052371f9ee81f31836ca27f577d195156aabb59a580d0241c1e2456309dd2a80"} Dec 03 16:26:34 crc kubenswrapper[4998]: I1203 16:26:34.218978 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17475de4-e997-47c0-83be-9a6f9e5f9857","Type":"ContainerStarted","Data":"1eaacd0fbcd3cee82668d38ad2e84204564f06d26ebed1e1a46cfd0f7da0272f"} Dec 03 16:26:36 crc kubenswrapper[4998]: I1203 16:26:36.245181 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17475de4-e997-47c0-83be-9a6f9e5f9857","Type":"ContainerStarted","Data":"d37e014a941ddda7c362d919d73551a49a4e220d0ed74dcc25d6a822ba8ac1a6"} Dec 03 16:26:36 crc kubenswrapper[4998]: I1203 16:26:36.246000 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 16:26:36 crc kubenswrapper[4998]: I1203 16:26:36.282448 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.743254005 podStartE2EDuration="6.282418255s" podCreationTimestamp="2025-12-03 16:26:30 +0000 UTC" firstStartedPulling="2025-12-03 16:26:31.104554917 +0000 UTC m=+1369.716255160" lastFinishedPulling="2025-12-03 16:26:35.643719187 +0000 UTC m=+1374.255419410" observedRunningTime="2025-12-03 16:26:36.271380343 +0000 UTC m=+1374.883080566" watchObservedRunningTime="2025-12-03 16:26:36.282418255 +0000 UTC m=+1374.894118478" Dec 03 16:26:40 crc kubenswrapper[4998]: I1203 16:26:40.785442 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:26:40 crc kubenswrapper[4998]: I1203 16:26:40.786293 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="17475de4-e997-47c0-83be-9a6f9e5f9857" containerName="ceilometer-central-agent" containerID="cri-o://052371f9ee81f31836ca27f577d195156aabb59a580d0241c1e2456309dd2a80" gracePeriod=30 Dec 03 16:26:40 crc kubenswrapper[4998]: I1203 16:26:40.786373 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="17475de4-e997-47c0-83be-9a6f9e5f9857" containerName="sg-core" containerID="cri-o://1eaacd0fbcd3cee82668d38ad2e84204564f06d26ebed1e1a46cfd0f7da0272f" gracePeriod=30 Dec 03 16:26:40 crc kubenswrapper[4998]: I1203 16:26:40.786415 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="17475de4-e997-47c0-83be-9a6f9e5f9857" containerName="ceilometer-notification-agent" containerID="cri-o://9d370332bd21a0f3c8106952dc5a5162b551ed091e5b56a0624befd147c8cf34" gracePeriod=30 Dec 03 16:26:40 crc kubenswrapper[4998]: I1203 16:26:40.786512 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="17475de4-e997-47c0-83be-9a6f9e5f9857" containerName="proxy-httpd" containerID="cri-o://d37e014a941ddda7c362d919d73551a49a4e220d0ed74dcc25d6a822ba8ac1a6" gracePeriod=30 Dec 03 16:26:41 crc kubenswrapper[4998]: I1203 16:26:41.292832 4998 generic.go:334] "Generic (PLEG): container finished" podID="17475de4-e997-47c0-83be-9a6f9e5f9857" containerID="d37e014a941ddda7c362d919d73551a49a4e220d0ed74dcc25d6a822ba8ac1a6" exitCode=0 Dec 03 16:26:41 crc kubenswrapper[4998]: I1203 16:26:41.293256 4998 generic.go:334] "Generic (PLEG): container finished" podID="17475de4-e997-47c0-83be-9a6f9e5f9857" containerID="1eaacd0fbcd3cee82668d38ad2e84204564f06d26ebed1e1a46cfd0f7da0272f" exitCode=2 Dec 03 16:26:41 crc kubenswrapper[4998]: I1203 16:26:41.292932 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17475de4-e997-47c0-83be-9a6f9e5f9857","Type":"ContainerDied","Data":"d37e014a941ddda7c362d919d73551a49a4e220d0ed74dcc25d6a822ba8ac1a6"} Dec 03 16:26:41 crc kubenswrapper[4998]: I1203 16:26:41.293294 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17475de4-e997-47c0-83be-9a6f9e5f9857","Type":"ContainerDied","Data":"1eaacd0fbcd3cee82668d38ad2e84204564f06d26ebed1e1a46cfd0f7da0272f"} Dec 03 16:26:42 crc kubenswrapper[4998]: I1203 16:26:42.306271 4998 generic.go:334] "Generic (PLEG): container finished" podID="17475de4-e997-47c0-83be-9a6f9e5f9857" containerID="052371f9ee81f31836ca27f577d195156aabb59a580d0241c1e2456309dd2a80" exitCode=0 Dec 03 16:26:42 crc kubenswrapper[4998]: I1203 16:26:42.306325 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17475de4-e997-47c0-83be-9a6f9e5f9857","Type":"ContainerDied","Data":"052371f9ee81f31836ca27f577d195156aabb59a580d0241c1e2456309dd2a80"} Dec 03 16:26:44 crc kubenswrapper[4998]: I1203 16:26:44.338636 4998 generic.go:334] "Generic (PLEG): container finished" podID="17475de4-e997-47c0-83be-9a6f9e5f9857" containerID="9d370332bd21a0f3c8106952dc5a5162b551ed091e5b56a0624befd147c8cf34" exitCode=0 Dec 03 16:26:44 crc kubenswrapper[4998]: I1203 16:26:44.338703 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17475de4-e997-47c0-83be-9a6f9e5f9857","Type":"ContainerDied","Data":"9d370332bd21a0f3c8106952dc5a5162b551ed091e5b56a0624befd147c8cf34"} Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.113775 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.234101 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t68lc\" (UniqueName: \"kubernetes.io/projected/17475de4-e997-47c0-83be-9a6f9e5f9857-kube-api-access-t68lc\") pod \"17475de4-e997-47c0-83be-9a6f9e5f9857\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.234172 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17475de4-e997-47c0-83be-9a6f9e5f9857-config-data\") pod \"17475de4-e997-47c0-83be-9a6f9e5f9857\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.234328 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17475de4-e997-47c0-83be-9a6f9e5f9857-log-httpd\") pod \"17475de4-e997-47c0-83be-9a6f9e5f9857\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.234483 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17475de4-e997-47c0-83be-9a6f9e5f9857-scripts\") pod \"17475de4-e997-47c0-83be-9a6f9e5f9857\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.234509 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17475de4-e997-47c0-83be-9a6f9e5f9857-run-httpd\") pod \"17475de4-e997-47c0-83be-9a6f9e5f9857\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.234550 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17475de4-e997-47c0-83be-9a6f9e5f9857-combined-ca-bundle\") pod \"17475de4-e997-47c0-83be-9a6f9e5f9857\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.234596 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/17475de4-e997-47c0-83be-9a6f9e5f9857-sg-core-conf-yaml\") pod \"17475de4-e997-47c0-83be-9a6f9e5f9857\" (UID: \"17475de4-e997-47c0-83be-9a6f9e5f9857\") " Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.234931 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17475de4-e997-47c0-83be-9a6f9e5f9857-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "17475de4-e997-47c0-83be-9a6f9e5f9857" (UID: "17475de4-e997-47c0-83be-9a6f9e5f9857"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.235047 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17475de4-e997-47c0-83be-9a6f9e5f9857-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "17475de4-e997-47c0-83be-9a6f9e5f9857" (UID: "17475de4-e997-47c0-83be-9a6f9e5f9857"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.235286 4998 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17475de4-e997-47c0-83be-9a6f9e5f9857-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.235304 4998 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/17475de4-e997-47c0-83be-9a6f9e5f9857-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.247380 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17475de4-e997-47c0-83be-9a6f9e5f9857-scripts" (OuterVolumeSpecName: "scripts") pod "17475de4-e997-47c0-83be-9a6f9e5f9857" (UID: "17475de4-e997-47c0-83be-9a6f9e5f9857"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.248819 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17475de4-e997-47c0-83be-9a6f9e5f9857-kube-api-access-t68lc" (OuterVolumeSpecName: "kube-api-access-t68lc") pod "17475de4-e997-47c0-83be-9a6f9e5f9857" (UID: "17475de4-e997-47c0-83be-9a6f9e5f9857"). InnerVolumeSpecName "kube-api-access-t68lc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.261343 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17475de4-e997-47c0-83be-9a6f9e5f9857-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "17475de4-e997-47c0-83be-9a6f9e5f9857" (UID: "17475de4-e997-47c0-83be-9a6f9e5f9857"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.314304 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17475de4-e997-47c0-83be-9a6f9e5f9857-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "17475de4-e997-47c0-83be-9a6f9e5f9857" (UID: "17475de4-e997-47c0-83be-9a6f9e5f9857"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.332718 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17475de4-e997-47c0-83be-9a6f9e5f9857-config-data" (OuterVolumeSpecName: "config-data") pod "17475de4-e997-47c0-83be-9a6f9e5f9857" (UID: "17475de4-e997-47c0-83be-9a6f9e5f9857"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.339259 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t68lc\" (UniqueName: \"kubernetes.io/projected/17475de4-e997-47c0-83be-9a6f9e5f9857-kube-api-access-t68lc\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.339292 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17475de4-e997-47c0-83be-9a6f9e5f9857-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.339304 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17475de4-e997-47c0-83be-9a6f9e5f9857-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.339314 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17475de4-e997-47c0-83be-9a6f9e5f9857-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.339323 4998 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/17475de4-e997-47c0-83be-9a6f9e5f9857-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.352305 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"17475de4-e997-47c0-83be-9a6f9e5f9857","Type":"ContainerDied","Data":"ac3426c408f3e150921fdf71b2b1ed57b048a0ed7f8c4ff804a9d769378b274e"} Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.352359 4998 scope.go:117] "RemoveContainer" containerID="d37e014a941ddda7c362d919d73551a49a4e220d0ed74dcc25d6a822ba8ac1a6" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.352423 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.419056 4998 scope.go:117] "RemoveContainer" containerID="1eaacd0fbcd3cee82668d38ad2e84204564f06d26ebed1e1a46cfd0f7da0272f" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.430062 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.440828 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.453137 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.453428 4998 scope.go:117] "RemoveContainer" containerID="9d370332bd21a0f3c8106952dc5a5162b551ed091e5b56a0624befd147c8cf34" Dec 03 16:26:45 crc kubenswrapper[4998]: E1203 16:26:45.453679 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17475de4-e997-47c0-83be-9a6f9e5f9857" containerName="sg-core" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.453698 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="17475de4-e997-47c0-83be-9a6f9e5f9857" containerName="sg-core" Dec 03 16:26:45 crc kubenswrapper[4998]: E1203 16:26:45.453732 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17475de4-e997-47c0-83be-9a6f9e5f9857" containerName="ceilometer-notification-agent" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.453740 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="17475de4-e997-47c0-83be-9a6f9e5f9857" containerName="ceilometer-notification-agent" Dec 03 16:26:45 crc kubenswrapper[4998]: E1203 16:26:45.453783 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17475de4-e997-47c0-83be-9a6f9e5f9857" containerName="ceilometer-central-agent" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.453790 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="17475de4-e997-47c0-83be-9a6f9e5f9857" containerName="ceilometer-central-agent" Dec 03 16:26:45 crc kubenswrapper[4998]: E1203 16:26:45.453806 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17475de4-e997-47c0-83be-9a6f9e5f9857" containerName="proxy-httpd" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.453812 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="17475de4-e997-47c0-83be-9a6f9e5f9857" containerName="proxy-httpd" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.454047 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="17475de4-e997-47c0-83be-9a6f9e5f9857" containerName="proxy-httpd" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.454065 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="17475de4-e997-47c0-83be-9a6f9e5f9857" containerName="ceilometer-central-agent" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.454094 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="17475de4-e997-47c0-83be-9a6f9e5f9857" containerName="ceilometer-notification-agent" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.454102 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="17475de4-e997-47c0-83be-9a6f9e5f9857" containerName="sg-core" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.456253 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.460227 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.460434 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.472787 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.506615 4998 scope.go:117] "RemoveContainer" containerID="052371f9ee81f31836ca27f577d195156aabb59a580d0241c1e2456309dd2a80" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.543041 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac504098-7519-43e2-80ba-b3da6061efc4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " pod="openstack/ceilometer-0" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.543404 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac504098-7519-43e2-80ba-b3da6061efc4-scripts\") pod \"ceilometer-0\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " pod="openstack/ceilometer-0" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.543442 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df4bp\" (UniqueName: \"kubernetes.io/projected/ac504098-7519-43e2-80ba-b3da6061efc4-kube-api-access-df4bp\") pod \"ceilometer-0\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " pod="openstack/ceilometer-0" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.543475 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac504098-7519-43e2-80ba-b3da6061efc4-run-httpd\") pod \"ceilometer-0\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " pod="openstack/ceilometer-0" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.543497 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac504098-7519-43e2-80ba-b3da6061efc4-log-httpd\") pod \"ceilometer-0\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " pod="openstack/ceilometer-0" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.543531 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac504098-7519-43e2-80ba-b3da6061efc4-config-data\") pod \"ceilometer-0\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " pod="openstack/ceilometer-0" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.543558 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac504098-7519-43e2-80ba-b3da6061efc4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " pod="openstack/ceilometer-0" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.645062 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac504098-7519-43e2-80ba-b3da6061efc4-scripts\") pod \"ceilometer-0\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " pod="openstack/ceilometer-0" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.645111 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df4bp\" (UniqueName: \"kubernetes.io/projected/ac504098-7519-43e2-80ba-b3da6061efc4-kube-api-access-df4bp\") pod \"ceilometer-0\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " pod="openstack/ceilometer-0" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.645147 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac504098-7519-43e2-80ba-b3da6061efc4-run-httpd\") pod \"ceilometer-0\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " pod="openstack/ceilometer-0" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.645171 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac504098-7519-43e2-80ba-b3da6061efc4-log-httpd\") pod \"ceilometer-0\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " pod="openstack/ceilometer-0" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.645205 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac504098-7519-43e2-80ba-b3da6061efc4-config-data\") pod \"ceilometer-0\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " pod="openstack/ceilometer-0" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.645227 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac504098-7519-43e2-80ba-b3da6061efc4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " pod="openstack/ceilometer-0" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.645319 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac504098-7519-43e2-80ba-b3da6061efc4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " pod="openstack/ceilometer-0" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.646093 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac504098-7519-43e2-80ba-b3da6061efc4-log-httpd\") pod \"ceilometer-0\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " pod="openstack/ceilometer-0" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.646159 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac504098-7519-43e2-80ba-b3da6061efc4-run-httpd\") pod \"ceilometer-0\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " pod="openstack/ceilometer-0" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.649425 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac504098-7519-43e2-80ba-b3da6061efc4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " pod="openstack/ceilometer-0" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.649711 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac504098-7519-43e2-80ba-b3da6061efc4-scripts\") pod \"ceilometer-0\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " pod="openstack/ceilometer-0" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.649976 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac504098-7519-43e2-80ba-b3da6061efc4-config-data\") pod \"ceilometer-0\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " pod="openstack/ceilometer-0" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.651424 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac504098-7519-43e2-80ba-b3da6061efc4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " pod="openstack/ceilometer-0" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.662798 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df4bp\" (UniqueName: \"kubernetes.io/projected/ac504098-7519-43e2-80ba-b3da6061efc4-kube-api-access-df4bp\") pod \"ceilometer-0\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " pod="openstack/ceilometer-0" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.688250 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17475de4-e997-47c0-83be-9a6f9e5f9857" path="/var/lib/kubelet/pods/17475de4-e997-47c0-83be-9a6f9e5f9857/volumes" Dec 03 16:26:45 crc kubenswrapper[4998]: I1203 16:26:45.791533 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:26:46 crc kubenswrapper[4998]: I1203 16:26:46.289091 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:26:46 crc kubenswrapper[4998]: W1203 16:26:46.292864 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac504098_7519_43e2_80ba_b3da6061efc4.slice/crio-3ec0705c5c1d063e67a3907edad29bc88b6cb739a6e44348cbe69518694a2618 WatchSource:0}: Error finding container 3ec0705c5c1d063e67a3907edad29bc88b6cb739a6e44348cbe69518694a2618: Status 404 returned error can't find the container with id 3ec0705c5c1d063e67a3907edad29bc88b6cb739a6e44348cbe69518694a2618 Dec 03 16:26:46 crc kubenswrapper[4998]: I1203 16:26:46.366254 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac504098-7519-43e2-80ba-b3da6061efc4","Type":"ContainerStarted","Data":"3ec0705c5c1d063e67a3907edad29bc88b6cb739a6e44348cbe69518694a2618"} Dec 03 16:26:46 crc kubenswrapper[4998]: I1203 16:26:46.372998 4998 generic.go:334] "Generic (PLEG): container finished" podID="3ff4ba4e-9b43-41fd-902b-977b9cc3aba9" containerID="c4fe18fac933af612bc197d41649650f29ab6e8496268bd2342d1c84935c7234" exitCode=0 Dec 03 16:26:46 crc kubenswrapper[4998]: I1203 16:26:46.373048 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6zkx5" event={"ID":"3ff4ba4e-9b43-41fd-902b-977b9cc3aba9","Type":"ContainerDied","Data":"c4fe18fac933af612bc197d41649650f29ab6e8496268bd2342d1c84935c7234"} Dec 03 16:26:47 crc kubenswrapper[4998]: I1203 16:26:47.384325 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac504098-7519-43e2-80ba-b3da6061efc4","Type":"ContainerStarted","Data":"264ea4221f3a7fdc459a2d05f76fa01f2aba5f42bd13e4c928869da5d4cf8838"} Dec 03 16:26:47 crc kubenswrapper[4998]: I1203 16:26:47.384988 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac504098-7519-43e2-80ba-b3da6061efc4","Type":"ContainerStarted","Data":"b5a5d7ac8282290b1e195561c9e717c4160d4d017b164b09665421b9d7d39472"} Dec 03 16:26:47 crc kubenswrapper[4998]: I1203 16:26:47.768147 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6zkx5" Dec 03 16:26:47 crc kubenswrapper[4998]: I1203 16:26:47.905374 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ff4ba4e-9b43-41fd-902b-977b9cc3aba9-combined-ca-bundle\") pod \"3ff4ba4e-9b43-41fd-902b-977b9cc3aba9\" (UID: \"3ff4ba4e-9b43-41fd-902b-977b9cc3aba9\") " Dec 03 16:26:47 crc kubenswrapper[4998]: I1203 16:26:47.905501 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ff4ba4e-9b43-41fd-902b-977b9cc3aba9-scripts\") pod \"3ff4ba4e-9b43-41fd-902b-977b9cc3aba9\" (UID: \"3ff4ba4e-9b43-41fd-902b-977b9cc3aba9\") " Dec 03 16:26:47 crc kubenswrapper[4998]: I1203 16:26:47.905539 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lbmx\" (UniqueName: \"kubernetes.io/projected/3ff4ba4e-9b43-41fd-902b-977b9cc3aba9-kube-api-access-4lbmx\") pod \"3ff4ba4e-9b43-41fd-902b-977b9cc3aba9\" (UID: \"3ff4ba4e-9b43-41fd-902b-977b9cc3aba9\") " Dec 03 16:26:47 crc kubenswrapper[4998]: I1203 16:26:47.905652 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ff4ba4e-9b43-41fd-902b-977b9cc3aba9-config-data\") pod \"3ff4ba4e-9b43-41fd-902b-977b9cc3aba9\" (UID: \"3ff4ba4e-9b43-41fd-902b-977b9cc3aba9\") " Dec 03 16:26:47 crc kubenswrapper[4998]: I1203 16:26:47.911678 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ff4ba4e-9b43-41fd-902b-977b9cc3aba9-scripts" (OuterVolumeSpecName: "scripts") pod "3ff4ba4e-9b43-41fd-902b-977b9cc3aba9" (UID: "3ff4ba4e-9b43-41fd-902b-977b9cc3aba9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:47 crc kubenswrapper[4998]: I1203 16:26:47.912500 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ff4ba4e-9b43-41fd-902b-977b9cc3aba9-kube-api-access-4lbmx" (OuterVolumeSpecName: "kube-api-access-4lbmx") pod "3ff4ba4e-9b43-41fd-902b-977b9cc3aba9" (UID: "3ff4ba4e-9b43-41fd-902b-977b9cc3aba9"). InnerVolumeSpecName "kube-api-access-4lbmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:26:47 crc kubenswrapper[4998]: I1203 16:26:47.932557 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ff4ba4e-9b43-41fd-902b-977b9cc3aba9-config-data" (OuterVolumeSpecName: "config-data") pod "3ff4ba4e-9b43-41fd-902b-977b9cc3aba9" (UID: "3ff4ba4e-9b43-41fd-902b-977b9cc3aba9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:47 crc kubenswrapper[4998]: I1203 16:26:47.941571 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ff4ba4e-9b43-41fd-902b-977b9cc3aba9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ff4ba4e-9b43-41fd-902b-977b9cc3aba9" (UID: "3ff4ba4e-9b43-41fd-902b-977b9cc3aba9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:26:48 crc kubenswrapper[4998]: I1203 16:26:48.007684 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ff4ba4e-9b43-41fd-902b-977b9cc3aba9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:48 crc kubenswrapper[4998]: I1203 16:26:48.007719 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ff4ba4e-9b43-41fd-902b-977b9cc3aba9-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:48 crc kubenswrapper[4998]: I1203 16:26:48.007730 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lbmx\" (UniqueName: \"kubernetes.io/projected/3ff4ba4e-9b43-41fd-902b-977b9cc3aba9-kube-api-access-4lbmx\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:48 crc kubenswrapper[4998]: I1203 16:26:48.007740 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ff4ba4e-9b43-41fd-902b-977b9cc3aba9-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:26:48 crc kubenswrapper[4998]: I1203 16:26:48.397398 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-6zkx5" event={"ID":"3ff4ba4e-9b43-41fd-902b-977b9cc3aba9","Type":"ContainerDied","Data":"2f4fd4359a332b21964064a20bc7de44fd4fd1822fdefccedc0cb1dab703e357"} Dec 03 16:26:48 crc kubenswrapper[4998]: I1203 16:26:48.397458 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f4fd4359a332b21964064a20bc7de44fd4fd1822fdefccedc0cb1dab703e357" Dec 03 16:26:48 crc kubenswrapper[4998]: I1203 16:26:48.397535 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-6zkx5" Dec 03 16:26:48 crc kubenswrapper[4998]: I1203 16:26:48.405957 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac504098-7519-43e2-80ba-b3da6061efc4","Type":"ContainerStarted","Data":"45258dafb53d8eba634abe2f783c47b774b6165cbb5a248abea8ae51dd949cbb"} Dec 03 16:26:48 crc kubenswrapper[4998]: I1203 16:26:48.482178 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 16:26:48 crc kubenswrapper[4998]: E1203 16:26:48.482727 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ff4ba4e-9b43-41fd-902b-977b9cc3aba9" containerName="nova-cell0-conductor-db-sync" Dec 03 16:26:48 crc kubenswrapper[4998]: I1203 16:26:48.482747 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ff4ba4e-9b43-41fd-902b-977b9cc3aba9" containerName="nova-cell0-conductor-db-sync" Dec 03 16:26:48 crc kubenswrapper[4998]: I1203 16:26:48.483007 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ff4ba4e-9b43-41fd-902b-977b9cc3aba9" containerName="nova-cell0-conductor-db-sync" Dec 03 16:26:48 crc kubenswrapper[4998]: I1203 16:26:48.483856 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 16:26:48 crc kubenswrapper[4998]: I1203 16:26:48.487069 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 16:26:48 crc kubenswrapper[4998]: I1203 16:26:48.487290 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-xcs6z" Dec 03 16:26:48 crc kubenswrapper[4998]: I1203 16:26:48.506827 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 16:26:48 crc kubenswrapper[4998]: I1203 16:26:48.619993 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97d1fa21-8374-4329-ac07-27d18803c2ab-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"97d1fa21-8374-4329-ac07-27d18803c2ab\") " pod="openstack/nova-cell0-conductor-0" Dec 03 16:26:48 crc kubenswrapper[4998]: I1203 16:26:48.620053 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dk7d\" (UniqueName: \"kubernetes.io/projected/97d1fa21-8374-4329-ac07-27d18803c2ab-kube-api-access-2dk7d\") pod \"nova-cell0-conductor-0\" (UID: \"97d1fa21-8374-4329-ac07-27d18803c2ab\") " pod="openstack/nova-cell0-conductor-0" Dec 03 16:26:48 crc kubenswrapper[4998]: I1203 16:26:48.620096 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97d1fa21-8374-4329-ac07-27d18803c2ab-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"97d1fa21-8374-4329-ac07-27d18803c2ab\") " pod="openstack/nova-cell0-conductor-0" Dec 03 16:26:48 crc kubenswrapper[4998]: I1203 16:26:48.721845 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97d1fa21-8374-4329-ac07-27d18803c2ab-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"97d1fa21-8374-4329-ac07-27d18803c2ab\") " pod="openstack/nova-cell0-conductor-0" Dec 03 16:26:48 crc kubenswrapper[4998]: I1203 16:26:48.721904 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dk7d\" (UniqueName: \"kubernetes.io/projected/97d1fa21-8374-4329-ac07-27d18803c2ab-kube-api-access-2dk7d\") pod \"nova-cell0-conductor-0\" (UID: \"97d1fa21-8374-4329-ac07-27d18803c2ab\") " pod="openstack/nova-cell0-conductor-0" Dec 03 16:26:48 crc kubenswrapper[4998]: I1203 16:26:48.721942 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97d1fa21-8374-4329-ac07-27d18803c2ab-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"97d1fa21-8374-4329-ac07-27d18803c2ab\") " pod="openstack/nova-cell0-conductor-0" Dec 03 16:26:48 crc kubenswrapper[4998]: I1203 16:26:48.725611 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97d1fa21-8374-4329-ac07-27d18803c2ab-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"97d1fa21-8374-4329-ac07-27d18803c2ab\") " pod="openstack/nova-cell0-conductor-0" Dec 03 16:26:48 crc kubenswrapper[4998]: I1203 16:26:48.726222 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97d1fa21-8374-4329-ac07-27d18803c2ab-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"97d1fa21-8374-4329-ac07-27d18803c2ab\") " pod="openstack/nova-cell0-conductor-0" Dec 03 16:26:48 crc kubenswrapper[4998]: I1203 16:26:48.759360 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dk7d\" (UniqueName: \"kubernetes.io/projected/97d1fa21-8374-4329-ac07-27d18803c2ab-kube-api-access-2dk7d\") pod \"nova-cell0-conductor-0\" (UID: \"97d1fa21-8374-4329-ac07-27d18803c2ab\") " pod="openstack/nova-cell0-conductor-0" Dec 03 16:26:48 crc kubenswrapper[4998]: I1203 16:26:48.802999 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 16:26:49 crc kubenswrapper[4998]: I1203 16:26:49.169816 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 16:26:49 crc kubenswrapper[4998]: I1203 16:26:49.415203 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"97d1fa21-8374-4329-ac07-27d18803c2ab","Type":"ContainerStarted","Data":"0ce442e2d96db3098133780303d67c1912d64c34027c8d1525712b32ddc35a15"} Dec 03 16:26:50 crc kubenswrapper[4998]: I1203 16:26:50.437907 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac504098-7519-43e2-80ba-b3da6061efc4","Type":"ContainerStarted","Data":"a586c5395556d7a52f9e901112e597939bedb0c521f42bc9cbf5d4544587c489"} Dec 03 16:26:50 crc kubenswrapper[4998]: I1203 16:26:50.438857 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 16:26:50 crc kubenswrapper[4998]: I1203 16:26:50.443035 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"97d1fa21-8374-4329-ac07-27d18803c2ab","Type":"ContainerStarted","Data":"6a3df62788c24c5522401850aead224beace1ad4a293ecf583ac460c76f1f1b5"} Dec 03 16:26:50 crc kubenswrapper[4998]: I1203 16:26:50.443349 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 03 16:26:50 crc kubenswrapper[4998]: I1203 16:26:50.468406 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.486943244 podStartE2EDuration="5.468383802s" podCreationTimestamp="2025-12-03 16:26:45 +0000 UTC" firstStartedPulling="2025-12-03 16:26:46.296895097 +0000 UTC m=+1384.908595320" lastFinishedPulling="2025-12-03 16:26:49.278335655 +0000 UTC m=+1387.890035878" observedRunningTime="2025-12-03 16:26:50.466487254 +0000 UTC m=+1389.078187517" watchObservedRunningTime="2025-12-03 16:26:50.468383802 +0000 UTC m=+1389.080084055" Dec 03 16:26:50 crc kubenswrapper[4998]: I1203 16:26:50.502441 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.502420941 podStartE2EDuration="2.502420941s" podCreationTimestamp="2025-12-03 16:26:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:26:50.497630081 +0000 UTC m=+1389.109330304" watchObservedRunningTime="2025-12-03 16:26:50.502420941 +0000 UTC m=+1389.114121164" Dec 03 16:26:58 crc kubenswrapper[4998]: I1203 16:26:58.838413 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.399244 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-jzsvl"] Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.400616 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-jzsvl" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.409659 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.409821 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.410411 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-jzsvl"] Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.518684 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.524485 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.526657 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.536816 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.543796 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7z7ls\" (UniqueName: \"kubernetes.io/projected/0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d-kube-api-access-7z7ls\") pod \"nova-cell0-cell-mapping-jzsvl\" (UID: \"0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d\") " pod="openstack/nova-cell0-cell-mapping-jzsvl" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.543860 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9c25919-a65d-4e03-93e1-4d19a42b3d28-config-data\") pod \"nova-api-0\" (UID: \"f9c25919-a65d-4e03-93e1-4d19a42b3d28\") " pod="openstack/nova-api-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.543904 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9c25919-a65d-4e03-93e1-4d19a42b3d28-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f9c25919-a65d-4e03-93e1-4d19a42b3d28\") " pod="openstack/nova-api-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.543925 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d-config-data\") pod \"nova-cell0-cell-mapping-jzsvl\" (UID: \"0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d\") " pod="openstack/nova-cell0-cell-mapping-jzsvl" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.543970 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9c25919-a65d-4e03-93e1-4d19a42b3d28-logs\") pod \"nova-api-0\" (UID: \"f9c25919-a65d-4e03-93e1-4d19a42b3d28\") " pod="openstack/nova-api-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.544005 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8grrm\" (UniqueName: \"kubernetes.io/projected/f9c25919-a65d-4e03-93e1-4d19a42b3d28-kube-api-access-8grrm\") pod \"nova-api-0\" (UID: \"f9c25919-a65d-4e03-93e1-4d19a42b3d28\") " pod="openstack/nova-api-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.544037 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-jzsvl\" (UID: \"0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d\") " pod="openstack/nova-cell0-cell-mapping-jzsvl" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.544055 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d-scripts\") pod \"nova-cell0-cell-mapping-jzsvl\" (UID: \"0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d\") " pod="openstack/nova-cell0-cell-mapping-jzsvl" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.578554 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.579742 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.587218 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.594188 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.644775 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-jzsvl\" (UID: \"0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d\") " pod="openstack/nova-cell0-cell-mapping-jzsvl" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.644814 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d-scripts\") pod \"nova-cell0-cell-mapping-jzsvl\" (UID: \"0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d\") " pod="openstack/nova-cell0-cell-mapping-jzsvl" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.644863 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7z7ls\" (UniqueName: \"kubernetes.io/projected/0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d-kube-api-access-7z7ls\") pod \"nova-cell0-cell-mapping-jzsvl\" (UID: \"0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d\") " pod="openstack/nova-cell0-cell-mapping-jzsvl" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.644894 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9c25919-a65d-4e03-93e1-4d19a42b3d28-config-data\") pod \"nova-api-0\" (UID: \"f9c25919-a65d-4e03-93e1-4d19a42b3d28\") " pod="openstack/nova-api-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.644935 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9c25919-a65d-4e03-93e1-4d19a42b3d28-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f9c25919-a65d-4e03-93e1-4d19a42b3d28\") " pod="openstack/nova-api-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.644953 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d-config-data\") pod \"nova-cell0-cell-mapping-jzsvl\" (UID: \"0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d\") " pod="openstack/nova-cell0-cell-mapping-jzsvl" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.644979 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37109c33-9436-40d5-aa8f-bb7e7c3a0f83-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"37109c33-9436-40d5-aa8f-bb7e7c3a0f83\") " pod="openstack/nova-scheduler-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.645002 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37109c33-9436-40d5-aa8f-bb7e7c3a0f83-config-data\") pod \"nova-scheduler-0\" (UID: \"37109c33-9436-40d5-aa8f-bb7e7c3a0f83\") " pod="openstack/nova-scheduler-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.645020 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lglp\" (UniqueName: \"kubernetes.io/projected/37109c33-9436-40d5-aa8f-bb7e7c3a0f83-kube-api-access-8lglp\") pod \"nova-scheduler-0\" (UID: \"37109c33-9436-40d5-aa8f-bb7e7c3a0f83\") " pod="openstack/nova-scheduler-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.645053 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9c25919-a65d-4e03-93e1-4d19a42b3d28-logs\") pod \"nova-api-0\" (UID: \"f9c25919-a65d-4e03-93e1-4d19a42b3d28\") " pod="openstack/nova-api-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.645087 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8grrm\" (UniqueName: \"kubernetes.io/projected/f9c25919-a65d-4e03-93e1-4d19a42b3d28-kube-api-access-8grrm\") pod \"nova-api-0\" (UID: \"f9c25919-a65d-4e03-93e1-4d19a42b3d28\") " pod="openstack/nova-api-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.648337 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9c25919-a65d-4e03-93e1-4d19a42b3d28-logs\") pod \"nova-api-0\" (UID: \"f9c25919-a65d-4e03-93e1-4d19a42b3d28\") " pod="openstack/nova-api-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.658924 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d-scripts\") pod \"nova-cell0-cell-mapping-jzsvl\" (UID: \"0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d\") " pod="openstack/nova-cell0-cell-mapping-jzsvl" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.658948 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9c25919-a65d-4e03-93e1-4d19a42b3d28-config-data\") pod \"nova-api-0\" (UID: \"f9c25919-a65d-4e03-93e1-4d19a42b3d28\") " pod="openstack/nova-api-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.660168 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d-config-data\") pod \"nova-cell0-cell-mapping-jzsvl\" (UID: \"0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d\") " pod="openstack/nova-cell0-cell-mapping-jzsvl" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.660384 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9c25919-a65d-4e03-93e1-4d19a42b3d28-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f9c25919-a65d-4e03-93e1-4d19a42b3d28\") " pod="openstack/nova-api-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.662216 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.663456 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.667085 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.681952 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8grrm\" (UniqueName: \"kubernetes.io/projected/f9c25919-a65d-4e03-93e1-4d19a42b3d28-kube-api-access-8grrm\") pod \"nova-api-0\" (UID: \"f9c25919-a65d-4e03-93e1-4d19a42b3d28\") " pod="openstack/nova-api-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.696404 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-jzsvl\" (UID: \"0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d\") " pod="openstack/nova-cell0-cell-mapping-jzsvl" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.713869 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.714966 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7z7ls\" (UniqueName: \"kubernetes.io/projected/0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d-kube-api-access-7z7ls\") pod \"nova-cell0-cell-mapping-jzsvl\" (UID: \"0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d\") " pod="openstack/nova-cell0-cell-mapping-jzsvl" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.721212 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.739892 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-jzsvl" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.742285 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.751017 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.752023 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a4df5d0-e6f7-403b-898a-46d3240f9100-config-data\") pod \"nova-metadata-0\" (UID: \"1a4df5d0-e6f7-403b-898a-46d3240f9100\") " pod="openstack/nova-metadata-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.752168 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cfaa0aa-a7ff-4349-ae1f-2403f81041b0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6cfaa0aa-a7ff-4349-ae1f-2403f81041b0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.752218 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cfaa0aa-a7ff-4349-ae1f-2403f81041b0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6cfaa0aa-a7ff-4349-ae1f-2403f81041b0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.752233 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a4df5d0-e6f7-403b-898a-46d3240f9100-logs\") pod \"nova-metadata-0\" (UID: \"1a4df5d0-e6f7-403b-898a-46d3240f9100\") " pod="openstack/nova-metadata-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.752283 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37109c33-9436-40d5-aa8f-bb7e7c3a0f83-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"37109c33-9436-40d5-aa8f-bb7e7c3a0f83\") " pod="openstack/nova-scheduler-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.752323 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37109c33-9436-40d5-aa8f-bb7e7c3a0f83-config-data\") pod \"nova-scheduler-0\" (UID: \"37109c33-9436-40d5-aa8f-bb7e7c3a0f83\") " pod="openstack/nova-scheduler-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.752344 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lglp\" (UniqueName: \"kubernetes.io/projected/37109c33-9436-40d5-aa8f-bb7e7c3a0f83-kube-api-access-8lglp\") pod \"nova-scheduler-0\" (UID: \"37109c33-9436-40d5-aa8f-bb7e7c3a0f83\") " pod="openstack/nova-scheduler-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.752381 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m7rw\" (UniqueName: \"kubernetes.io/projected/6cfaa0aa-a7ff-4349-ae1f-2403f81041b0-kube-api-access-2m7rw\") pod \"nova-cell1-novncproxy-0\" (UID: \"6cfaa0aa-a7ff-4349-ae1f-2403f81041b0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.752410 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blm22\" (UniqueName: \"kubernetes.io/projected/1a4df5d0-e6f7-403b-898a-46d3240f9100-kube-api-access-blm22\") pod \"nova-metadata-0\" (UID: \"1a4df5d0-e6f7-403b-898a-46d3240f9100\") " pod="openstack/nova-metadata-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.752581 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a4df5d0-e6f7-403b-898a-46d3240f9100-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1a4df5d0-e6f7-403b-898a-46d3240f9100\") " pod="openstack/nova-metadata-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.760239 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37109c33-9436-40d5-aa8f-bb7e7c3a0f83-config-data\") pod \"nova-scheduler-0\" (UID: \"37109c33-9436-40d5-aa8f-bb7e7c3a0f83\") " pod="openstack/nova-scheduler-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.776505 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.798521 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37109c33-9436-40d5-aa8f-bb7e7c3a0f83-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"37109c33-9436-40d5-aa8f-bb7e7c3a0f83\") " pod="openstack/nova-scheduler-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.813035 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lglp\" (UniqueName: \"kubernetes.io/projected/37109c33-9436-40d5-aa8f-bb7e7c3a0f83-kube-api-access-8lglp\") pod \"nova-scheduler-0\" (UID: \"37109c33-9436-40d5-aa8f-bb7e7c3a0f83\") " pod="openstack/nova-scheduler-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.853581 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cfaa0aa-a7ff-4349-ae1f-2403f81041b0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6cfaa0aa-a7ff-4349-ae1f-2403f81041b0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.853624 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a4df5d0-e6f7-403b-898a-46d3240f9100-logs\") pod \"nova-metadata-0\" (UID: \"1a4df5d0-e6f7-403b-898a-46d3240f9100\") " pod="openstack/nova-metadata-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.853784 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m7rw\" (UniqueName: \"kubernetes.io/projected/6cfaa0aa-a7ff-4349-ae1f-2403f81041b0-kube-api-access-2m7rw\") pod \"nova-cell1-novncproxy-0\" (UID: \"6cfaa0aa-a7ff-4349-ae1f-2403f81041b0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.853808 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blm22\" (UniqueName: \"kubernetes.io/projected/1a4df5d0-e6f7-403b-898a-46d3240f9100-kube-api-access-blm22\") pod \"nova-metadata-0\" (UID: \"1a4df5d0-e6f7-403b-898a-46d3240f9100\") " pod="openstack/nova-metadata-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.853883 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a4df5d0-e6f7-403b-898a-46d3240f9100-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1a4df5d0-e6f7-403b-898a-46d3240f9100\") " pod="openstack/nova-metadata-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.853904 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a4df5d0-e6f7-403b-898a-46d3240f9100-config-data\") pod \"nova-metadata-0\" (UID: \"1a4df5d0-e6f7-403b-898a-46d3240f9100\") " pod="openstack/nova-metadata-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.853942 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cfaa0aa-a7ff-4349-ae1f-2403f81041b0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6cfaa0aa-a7ff-4349-ae1f-2403f81041b0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.856030 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a4df5d0-e6f7-403b-898a-46d3240f9100-logs\") pod \"nova-metadata-0\" (UID: \"1a4df5d0-e6f7-403b-898a-46d3240f9100\") " pod="openstack/nova-metadata-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.863244 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7995555d47-42cp9"] Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.865221 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7995555d47-42cp9" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.872451 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7995555d47-42cp9"] Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.877294 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a4df5d0-e6f7-403b-898a-46d3240f9100-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1a4df5d0-e6f7-403b-898a-46d3240f9100\") " pod="openstack/nova-metadata-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.878341 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cfaa0aa-a7ff-4349-ae1f-2403f81041b0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6cfaa0aa-a7ff-4349-ae1f-2403f81041b0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.880158 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a4df5d0-e6f7-403b-898a-46d3240f9100-config-data\") pod \"nova-metadata-0\" (UID: \"1a4df5d0-e6f7-403b-898a-46d3240f9100\") " pod="openstack/nova-metadata-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.880695 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.886151 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cfaa0aa-a7ff-4349-ae1f-2403f81041b0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6cfaa0aa-a7ff-4349-ae1f-2403f81041b0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.907542 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.914440 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blm22\" (UniqueName: \"kubernetes.io/projected/1a4df5d0-e6f7-403b-898a-46d3240f9100-kube-api-access-blm22\") pod \"nova-metadata-0\" (UID: \"1a4df5d0-e6f7-403b-898a-46d3240f9100\") " pod="openstack/nova-metadata-0" Dec 03 16:26:59 crc kubenswrapper[4998]: I1203 16:26:59.916590 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m7rw\" (UniqueName: \"kubernetes.io/projected/6cfaa0aa-a7ff-4349-ae1f-2403f81041b0-kube-api-access-2m7rw\") pod \"nova-cell1-novncproxy-0\" (UID: \"6cfaa0aa-a7ff-4349-ae1f-2403f81041b0\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.043532 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.060795 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-config\") pod \"dnsmasq-dns-7995555d47-42cp9\" (UID: \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\") " pod="openstack/dnsmasq-dns-7995555d47-42cp9" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.060842 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-ovsdbserver-nb\") pod \"dnsmasq-dns-7995555d47-42cp9\" (UID: \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\") " pod="openstack/dnsmasq-dns-7995555d47-42cp9" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.060918 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-dns-svc\") pod \"dnsmasq-dns-7995555d47-42cp9\" (UID: \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\") " pod="openstack/dnsmasq-dns-7995555d47-42cp9" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.060962 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhpck\" (UniqueName: \"kubernetes.io/projected/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-kube-api-access-vhpck\") pod \"dnsmasq-dns-7995555d47-42cp9\" (UID: \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\") " pod="openstack/dnsmasq-dns-7995555d47-42cp9" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.060993 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-ovsdbserver-sb\") pod \"dnsmasq-dns-7995555d47-42cp9\" (UID: \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\") " pod="openstack/dnsmasq-dns-7995555d47-42cp9" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.061017 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-dns-swift-storage-0\") pod \"dnsmasq-dns-7995555d47-42cp9\" (UID: \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\") " pod="openstack/dnsmasq-dns-7995555d47-42cp9" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.071738 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.163000 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-dns-swift-storage-0\") pod \"dnsmasq-dns-7995555d47-42cp9\" (UID: \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\") " pod="openstack/dnsmasq-dns-7995555d47-42cp9" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.163080 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-config\") pod \"dnsmasq-dns-7995555d47-42cp9\" (UID: \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\") " pod="openstack/dnsmasq-dns-7995555d47-42cp9" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.163118 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-ovsdbserver-nb\") pod \"dnsmasq-dns-7995555d47-42cp9\" (UID: \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\") " pod="openstack/dnsmasq-dns-7995555d47-42cp9" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.163199 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-dns-svc\") pod \"dnsmasq-dns-7995555d47-42cp9\" (UID: \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\") " pod="openstack/dnsmasq-dns-7995555d47-42cp9" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.163261 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhpck\" (UniqueName: \"kubernetes.io/projected/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-kube-api-access-vhpck\") pod \"dnsmasq-dns-7995555d47-42cp9\" (UID: \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\") " pod="openstack/dnsmasq-dns-7995555d47-42cp9" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.163292 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-ovsdbserver-sb\") pod \"dnsmasq-dns-7995555d47-42cp9\" (UID: \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\") " pod="openstack/dnsmasq-dns-7995555d47-42cp9" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.164056 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-ovsdbserver-sb\") pod \"dnsmasq-dns-7995555d47-42cp9\" (UID: \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\") " pod="openstack/dnsmasq-dns-7995555d47-42cp9" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.164402 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-dns-swift-storage-0\") pod \"dnsmasq-dns-7995555d47-42cp9\" (UID: \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\") " pod="openstack/dnsmasq-dns-7995555d47-42cp9" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.164797 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-dns-svc\") pod \"dnsmasq-dns-7995555d47-42cp9\" (UID: \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\") " pod="openstack/dnsmasq-dns-7995555d47-42cp9" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.165080 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-config\") pod \"dnsmasq-dns-7995555d47-42cp9\" (UID: \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\") " pod="openstack/dnsmasq-dns-7995555d47-42cp9" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.165600 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-ovsdbserver-nb\") pod \"dnsmasq-dns-7995555d47-42cp9\" (UID: \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\") " pod="openstack/dnsmasq-dns-7995555d47-42cp9" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.194304 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhpck\" (UniqueName: \"kubernetes.io/projected/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-kube-api-access-vhpck\") pod \"dnsmasq-dns-7995555d47-42cp9\" (UID: \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\") " pod="openstack/dnsmasq-dns-7995555d47-42cp9" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.398235 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7995555d47-42cp9" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.574640 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-jzsvl"] Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.579468 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-jzsvl" event={"ID":"0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d","Type":"ContainerStarted","Data":"954a0b7734993941d2aefbb239e007d250add17e72345ca486ca45fd68b68e0a"} Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.732573 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 16:27:00 crc kubenswrapper[4998]: W1203 16:27:00.737082 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9c25919_a65d_4e03_93e1_4d19a42b3d28.slice/crio-7e2d92aa5903e257688bb1e797558969d4984850a2aad26dd9dc787ba9bc89d7 WatchSource:0}: Error finding container 7e2d92aa5903e257688bb1e797558969d4984850a2aad26dd9dc787ba9bc89d7: Status 404 returned error can't find the container with id 7e2d92aa5903e257688bb1e797558969d4984850a2aad26dd9dc787ba9bc89d7 Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.813536 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.882245 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vvjng"] Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.883804 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vvjng" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.890048 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vvjng"] Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.890413 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.890554 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 03 16:27:00 crc kubenswrapper[4998]: W1203 16:27:00.971419 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6cfaa0aa_a7ff_4349_ae1f_2403f81041b0.slice/crio-21fbcbb11b108ab59c4a62b60991d51a461bc3ddd9e9731c16f9c82b3f670a06 WatchSource:0}: Error finding container 21fbcbb11b108ab59c4a62b60991d51a461bc3ddd9e9731c16f9c82b3f670a06: Status 404 returned error can't find the container with id 21fbcbb11b108ab59c4a62b60991d51a461bc3ddd9e9731c16f9c82b3f670a06 Dec 03 16:27:00 crc kubenswrapper[4998]: W1203 16:27:00.972680 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a4df5d0_e6f7_403b_898a_46d3240f9100.slice/crio-54e95b0aca21ab4b4fe5350b3017d05d603f9604bb70b7d38d6fe3b83af15fe1 WatchSource:0}: Error finding container 54e95b0aca21ab4b4fe5350b3017d05d603f9604bb70b7d38d6fe3b83af15fe1: Status 404 returned error can't find the container with id 54e95b0aca21ab4b4fe5350b3017d05d603f9604bb70b7d38d6fe3b83af15fe1 Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.981692 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.990085 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgtbd\" (UniqueName: \"kubernetes.io/projected/86d54b38-8913-4d18-92ee-490aceef1677-kube-api-access-vgtbd\") pod \"nova-cell1-conductor-db-sync-vvjng\" (UID: \"86d54b38-8913-4d18-92ee-490aceef1677\") " pod="openstack/nova-cell1-conductor-db-sync-vvjng" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.990159 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d54b38-8913-4d18-92ee-490aceef1677-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vvjng\" (UID: \"86d54b38-8913-4d18-92ee-490aceef1677\") " pod="openstack/nova-cell1-conductor-db-sync-vvjng" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.990318 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86d54b38-8913-4d18-92ee-490aceef1677-config-data\") pod \"nova-cell1-conductor-db-sync-vvjng\" (UID: \"86d54b38-8913-4d18-92ee-490aceef1677\") " pod="openstack/nova-cell1-conductor-db-sync-vvjng" Dec 03 16:27:00 crc kubenswrapper[4998]: I1203 16:27:00.990373 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86d54b38-8913-4d18-92ee-490aceef1677-scripts\") pod \"nova-cell1-conductor-db-sync-vvjng\" (UID: \"86d54b38-8913-4d18-92ee-490aceef1677\") " pod="openstack/nova-cell1-conductor-db-sync-vvjng" Dec 03 16:27:01 crc kubenswrapper[4998]: I1203 16:27:01.003573 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 16:27:01 crc kubenswrapper[4998]: I1203 16:27:01.092716 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d54b38-8913-4d18-92ee-490aceef1677-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vvjng\" (UID: \"86d54b38-8913-4d18-92ee-490aceef1677\") " pod="openstack/nova-cell1-conductor-db-sync-vvjng" Dec 03 16:27:01 crc kubenswrapper[4998]: I1203 16:27:01.092872 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86d54b38-8913-4d18-92ee-490aceef1677-config-data\") pod \"nova-cell1-conductor-db-sync-vvjng\" (UID: \"86d54b38-8913-4d18-92ee-490aceef1677\") " pod="openstack/nova-cell1-conductor-db-sync-vvjng" Dec 03 16:27:01 crc kubenswrapper[4998]: I1203 16:27:01.092929 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86d54b38-8913-4d18-92ee-490aceef1677-scripts\") pod \"nova-cell1-conductor-db-sync-vvjng\" (UID: \"86d54b38-8913-4d18-92ee-490aceef1677\") " pod="openstack/nova-cell1-conductor-db-sync-vvjng" Dec 03 16:27:01 crc kubenswrapper[4998]: I1203 16:27:01.092997 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgtbd\" (UniqueName: \"kubernetes.io/projected/86d54b38-8913-4d18-92ee-490aceef1677-kube-api-access-vgtbd\") pod \"nova-cell1-conductor-db-sync-vvjng\" (UID: \"86d54b38-8913-4d18-92ee-490aceef1677\") " pod="openstack/nova-cell1-conductor-db-sync-vvjng" Dec 03 16:27:01 crc kubenswrapper[4998]: I1203 16:27:01.098110 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86d54b38-8913-4d18-92ee-490aceef1677-scripts\") pod \"nova-cell1-conductor-db-sync-vvjng\" (UID: \"86d54b38-8913-4d18-92ee-490aceef1677\") " pod="openstack/nova-cell1-conductor-db-sync-vvjng" Dec 03 16:27:01 crc kubenswrapper[4998]: I1203 16:27:01.100441 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86d54b38-8913-4d18-92ee-490aceef1677-config-data\") pod \"nova-cell1-conductor-db-sync-vvjng\" (UID: \"86d54b38-8913-4d18-92ee-490aceef1677\") " pod="openstack/nova-cell1-conductor-db-sync-vvjng" Dec 03 16:27:01 crc kubenswrapper[4998]: I1203 16:27:01.103850 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d54b38-8913-4d18-92ee-490aceef1677-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vvjng\" (UID: \"86d54b38-8913-4d18-92ee-490aceef1677\") " pod="openstack/nova-cell1-conductor-db-sync-vvjng" Dec 03 16:27:01 crc kubenswrapper[4998]: I1203 16:27:01.113920 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgtbd\" (UniqueName: \"kubernetes.io/projected/86d54b38-8913-4d18-92ee-490aceef1677-kube-api-access-vgtbd\") pod \"nova-cell1-conductor-db-sync-vvjng\" (UID: \"86d54b38-8913-4d18-92ee-490aceef1677\") " pod="openstack/nova-cell1-conductor-db-sync-vvjng" Dec 03 16:27:01 crc kubenswrapper[4998]: I1203 16:27:01.197882 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7995555d47-42cp9"] Dec 03 16:27:01 crc kubenswrapper[4998]: I1203 16:27:01.209450 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vvjng" Dec 03 16:27:01 crc kubenswrapper[4998]: W1203 16:27:01.212034 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc435b8fa_d8c6_49fb_b339_edeb6a3f182c.slice/crio-1b71958fe17dab436da10becd2e5433e45b3b888e7dd2bcac07ecfcfb8a28e51 WatchSource:0}: Error finding container 1b71958fe17dab436da10becd2e5433e45b3b888e7dd2bcac07ecfcfb8a28e51: Status 404 returned error can't find the container with id 1b71958fe17dab436da10becd2e5433e45b3b888e7dd2bcac07ecfcfb8a28e51 Dec 03 16:27:01 crc kubenswrapper[4998]: I1203 16:27:01.598719 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1a4df5d0-e6f7-403b-898a-46d3240f9100","Type":"ContainerStarted","Data":"54e95b0aca21ab4b4fe5350b3017d05d603f9604bb70b7d38d6fe3b83af15fe1"} Dec 03 16:27:01 crc kubenswrapper[4998]: I1203 16:27:01.601590 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-jzsvl" event={"ID":"0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d","Type":"ContainerStarted","Data":"e9f429e5e1f1bd89d2bb9cca65fdc0eebb511a8d993f432a17e22b665be35a4c"} Dec 03 16:27:01 crc kubenswrapper[4998]: I1203 16:27:01.603649 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"37109c33-9436-40d5-aa8f-bb7e7c3a0f83","Type":"ContainerStarted","Data":"2bd0a3caf659a72d355c785d421298283ee9212cb2d4663664122fea120c43a0"} Dec 03 16:27:01 crc kubenswrapper[4998]: I1203 16:27:01.610056 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6cfaa0aa-a7ff-4349-ae1f-2403f81041b0","Type":"ContainerStarted","Data":"21fbcbb11b108ab59c4a62b60991d51a461bc3ddd9e9731c16f9c82b3f670a06"} Dec 03 16:27:01 crc kubenswrapper[4998]: I1203 16:27:01.613097 4998 generic.go:334] "Generic (PLEG): container finished" podID="c435b8fa-d8c6-49fb-b339-edeb6a3f182c" containerID="1799f2ac959a3ee61295e4179dcd841540aad9731c9842955db0dbccd29b29b9" exitCode=0 Dec 03 16:27:01 crc kubenswrapper[4998]: I1203 16:27:01.613162 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7995555d47-42cp9" event={"ID":"c435b8fa-d8c6-49fb-b339-edeb6a3f182c","Type":"ContainerDied","Data":"1799f2ac959a3ee61295e4179dcd841540aad9731c9842955db0dbccd29b29b9"} Dec 03 16:27:01 crc kubenswrapper[4998]: I1203 16:27:01.613190 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7995555d47-42cp9" event={"ID":"c435b8fa-d8c6-49fb-b339-edeb6a3f182c","Type":"ContainerStarted","Data":"1b71958fe17dab436da10becd2e5433e45b3b888e7dd2bcac07ecfcfb8a28e51"} Dec 03 16:27:01 crc kubenswrapper[4998]: I1203 16:27:01.617655 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-jzsvl" podStartSLOduration=2.61763695 podStartE2EDuration="2.61763695s" podCreationTimestamp="2025-12-03 16:26:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:27:01.616060847 +0000 UTC m=+1400.227761070" watchObservedRunningTime="2025-12-03 16:27:01.61763695 +0000 UTC m=+1400.229337173" Dec 03 16:27:01 crc kubenswrapper[4998]: I1203 16:27:01.618289 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f9c25919-a65d-4e03-93e1-4d19a42b3d28","Type":"ContainerStarted","Data":"7e2d92aa5903e257688bb1e797558969d4984850a2aad26dd9dc787ba9bc89d7"} Dec 03 16:27:01 crc kubenswrapper[4998]: I1203 16:27:01.792520 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vvjng"] Dec 03 16:27:02 crc kubenswrapper[4998]: I1203 16:27:02.628674 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vvjng" event={"ID":"86d54b38-8913-4d18-92ee-490aceef1677","Type":"ContainerStarted","Data":"a4321b32108f5fdfe6edbcb98b07cbdfb20c6f7fb29e9c891a769463a10e12b5"} Dec 03 16:27:03 crc kubenswrapper[4998]: I1203 16:27:03.637082 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 16:27:03 crc kubenswrapper[4998]: I1203 16:27:03.648353 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 16:27:04 crc kubenswrapper[4998]: I1203 16:27:04.648030 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vvjng" event={"ID":"86d54b38-8913-4d18-92ee-490aceef1677","Type":"ContainerStarted","Data":"ae7a2f32fd866e6275ac3717266de1857c58450ea3f2547815e726e96f085b84"} Dec 03 16:27:04 crc kubenswrapper[4998]: I1203 16:27:04.652826 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"37109c33-9436-40d5-aa8f-bb7e7c3a0f83","Type":"ContainerStarted","Data":"682763cd65cf4804dc534645f561db0b3daf020b18f1b16c9b083082fae8e47d"} Dec 03 16:27:04 crc kubenswrapper[4998]: I1203 16:27:04.655102 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6cfaa0aa-a7ff-4349-ae1f-2403f81041b0","Type":"ContainerStarted","Data":"f2306dc70408dbd35479757f2334c590768a431598426f2a974900cde2b177b7"} Dec 03 16:27:04 crc kubenswrapper[4998]: I1203 16:27:04.655212 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="6cfaa0aa-a7ff-4349-ae1f-2403f81041b0" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://f2306dc70408dbd35479757f2334c590768a431598426f2a974900cde2b177b7" gracePeriod=30 Dec 03 16:27:04 crc kubenswrapper[4998]: I1203 16:27:04.662989 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7995555d47-42cp9" event={"ID":"c435b8fa-d8c6-49fb-b339-edeb6a3f182c","Type":"ContainerStarted","Data":"c6719e5a27b5be3a5194c080d34db81c4c39ac724df7ddabfa3d9de6ce711dc9"} Dec 03 16:27:04 crc kubenswrapper[4998]: I1203 16:27:04.663789 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7995555d47-42cp9" Dec 03 16:27:04 crc kubenswrapper[4998]: I1203 16:27:04.669307 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f9c25919-a65d-4e03-93e1-4d19a42b3d28","Type":"ContainerStarted","Data":"c0df6738509fa9424f3e8f55f7f3a2cbfe198c9d93aa4df0fdf48d5497eea9f7"} Dec 03 16:27:04 crc kubenswrapper[4998]: I1203 16:27:04.669314 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-vvjng" podStartSLOduration=4.669296487 podStartE2EDuration="4.669296487s" podCreationTimestamp="2025-12-03 16:27:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:27:04.666249593 +0000 UTC m=+1403.277949826" watchObservedRunningTime="2025-12-03 16:27:04.669296487 +0000 UTC m=+1403.280996710" Dec 03 16:27:04 crc kubenswrapper[4998]: I1203 16:27:04.669350 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f9c25919-a65d-4e03-93e1-4d19a42b3d28","Type":"ContainerStarted","Data":"458902618fd0712551d5f6d404870bed957eecaa53187f738e42909f6f22cdd4"} Dec 03 16:27:04 crc kubenswrapper[4998]: I1203 16:27:04.671097 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1a4df5d0-e6f7-403b-898a-46d3240f9100","Type":"ContainerStarted","Data":"9184e20036fae6dea7fb09ef9673244bd0545bc4d1b9ef053b11aced17e6fe51"} Dec 03 16:27:04 crc kubenswrapper[4998]: I1203 16:27:04.671135 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1a4df5d0-e6f7-403b-898a-46d3240f9100","Type":"ContainerStarted","Data":"1d2cb0ab5309267aa896af5145b7f88c89ffdf464eaa0fe4029fd43bd43ed634"} Dec 03 16:27:04 crc kubenswrapper[4998]: I1203 16:27:04.671278 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1a4df5d0-e6f7-403b-898a-46d3240f9100" containerName="nova-metadata-log" containerID="cri-o://1d2cb0ab5309267aa896af5145b7f88c89ffdf464eaa0fe4029fd43bd43ed634" gracePeriod=30 Dec 03 16:27:04 crc kubenswrapper[4998]: I1203 16:27:04.671307 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1a4df5d0-e6f7-403b-898a-46d3240f9100" containerName="nova-metadata-metadata" containerID="cri-o://9184e20036fae6dea7fb09ef9673244bd0545bc4d1b9ef053b11aced17e6fe51" gracePeriod=30 Dec 03 16:27:04 crc kubenswrapper[4998]: I1203 16:27:04.680412 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.912283591 podStartE2EDuration="5.68039095s" podCreationTimestamp="2025-12-03 16:26:59 +0000 UTC" firstStartedPulling="2025-12-03 16:27:00.974215191 +0000 UTC m=+1399.585915404" lastFinishedPulling="2025-12-03 16:27:03.74232254 +0000 UTC m=+1402.354022763" observedRunningTime="2025-12-03 16:27:04.678559573 +0000 UTC m=+1403.290259806" watchObservedRunningTime="2025-12-03 16:27:04.68039095 +0000 UTC m=+1403.292091183" Dec 03 16:27:04 crc kubenswrapper[4998]: I1203 16:27:04.709570 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7995555d47-42cp9" podStartSLOduration=5.709551548 podStartE2EDuration="5.709551548s" podCreationTimestamp="2025-12-03 16:26:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:27:04.707139513 +0000 UTC m=+1403.318839736" watchObservedRunningTime="2025-12-03 16:27:04.709551548 +0000 UTC m=+1403.321251771" Dec 03 16:27:04 crc kubenswrapper[4998]: I1203 16:27:04.728333 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.816430396 podStartE2EDuration="5.728316193s" podCreationTimestamp="2025-12-03 16:26:59 +0000 UTC" firstStartedPulling="2025-12-03 16:27:00.814182824 +0000 UTC m=+1399.425883047" lastFinishedPulling="2025-12-03 16:27:03.726068621 +0000 UTC m=+1402.337768844" observedRunningTime="2025-12-03 16:27:04.722463677 +0000 UTC m=+1403.334163910" watchObservedRunningTime="2025-12-03 16:27:04.728316193 +0000 UTC m=+1403.340016416" Dec 03 16:27:04 crc kubenswrapper[4998]: I1203 16:27:04.810177 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.818189852 podStartE2EDuration="5.810156794s" podCreationTimestamp="2025-12-03 16:26:59 +0000 UTC" firstStartedPulling="2025-12-03 16:27:00.739935415 +0000 UTC m=+1399.351635638" lastFinishedPulling="2025-12-03 16:27:03.731902357 +0000 UTC m=+1402.343602580" observedRunningTime="2025-12-03 16:27:04.757119146 +0000 UTC m=+1403.368819369" watchObservedRunningTime="2025-12-03 16:27:04.810156794 +0000 UTC m=+1403.421857017" Dec 03 16:27:04 crc kubenswrapper[4998]: I1203 16:27:04.812263 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.05046703 podStartE2EDuration="5.812253745s" podCreationTimestamp="2025-12-03 16:26:59 +0000 UTC" firstStartedPulling="2025-12-03 16:27:00.980585835 +0000 UTC m=+1399.592286058" lastFinishedPulling="2025-12-03 16:27:03.74237255 +0000 UTC m=+1402.354072773" observedRunningTime="2025-12-03 16:27:04.775988243 +0000 UTC m=+1403.387688456" watchObservedRunningTime="2025-12-03 16:27:04.812253745 +0000 UTC m=+1403.423953968" Dec 03 16:27:04 crc kubenswrapper[4998]: I1203 16:27:04.908748 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 16:27:05 crc kubenswrapper[4998]: I1203 16:27:05.044494 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:05 crc kubenswrapper[4998]: I1203 16:27:05.072844 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 16:27:05 crc kubenswrapper[4998]: I1203 16:27:05.072901 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 16:27:05 crc kubenswrapper[4998]: I1203 16:27:05.687194 4998 generic.go:334] "Generic (PLEG): container finished" podID="1a4df5d0-e6f7-403b-898a-46d3240f9100" containerID="9184e20036fae6dea7fb09ef9673244bd0545bc4d1b9ef053b11aced17e6fe51" exitCode=0 Dec 03 16:27:05 crc kubenswrapper[4998]: I1203 16:27:05.687526 4998 generic.go:334] "Generic (PLEG): container finished" podID="1a4df5d0-e6f7-403b-898a-46d3240f9100" containerID="1d2cb0ab5309267aa896af5145b7f88c89ffdf464eaa0fe4029fd43bd43ed634" exitCode=143 Dec 03 16:27:05 crc kubenswrapper[4998]: I1203 16:27:05.699181 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1a4df5d0-e6f7-403b-898a-46d3240f9100","Type":"ContainerDied","Data":"9184e20036fae6dea7fb09ef9673244bd0545bc4d1b9ef053b11aced17e6fe51"} Dec 03 16:27:05 crc kubenswrapper[4998]: I1203 16:27:05.699223 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1a4df5d0-e6f7-403b-898a-46d3240f9100","Type":"ContainerDied","Data":"1d2cb0ab5309267aa896af5145b7f88c89ffdf464eaa0fe4029fd43bd43ed634"} Dec 03 16:27:05 crc kubenswrapper[4998]: I1203 16:27:05.699236 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1a4df5d0-e6f7-403b-898a-46d3240f9100","Type":"ContainerDied","Data":"54e95b0aca21ab4b4fe5350b3017d05d603f9604bb70b7d38d6fe3b83af15fe1"} Dec 03 16:27:05 crc kubenswrapper[4998]: I1203 16:27:05.699246 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54e95b0aca21ab4b4fe5350b3017d05d603f9604bb70b7d38d6fe3b83af15fe1" Dec 03 16:27:05 crc kubenswrapper[4998]: I1203 16:27:05.762195 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 16:27:05 crc kubenswrapper[4998]: I1203 16:27:05.913053 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a4df5d0-e6f7-403b-898a-46d3240f9100-logs\") pod \"1a4df5d0-e6f7-403b-898a-46d3240f9100\" (UID: \"1a4df5d0-e6f7-403b-898a-46d3240f9100\") " Dec 03 16:27:05 crc kubenswrapper[4998]: I1203 16:27:05.913139 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a4df5d0-e6f7-403b-898a-46d3240f9100-combined-ca-bundle\") pod \"1a4df5d0-e6f7-403b-898a-46d3240f9100\" (UID: \"1a4df5d0-e6f7-403b-898a-46d3240f9100\") " Dec 03 16:27:05 crc kubenswrapper[4998]: I1203 16:27:05.913246 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a4df5d0-e6f7-403b-898a-46d3240f9100-config-data\") pod \"1a4df5d0-e6f7-403b-898a-46d3240f9100\" (UID: \"1a4df5d0-e6f7-403b-898a-46d3240f9100\") " Dec 03 16:27:05 crc kubenswrapper[4998]: I1203 16:27:05.913284 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blm22\" (UniqueName: \"kubernetes.io/projected/1a4df5d0-e6f7-403b-898a-46d3240f9100-kube-api-access-blm22\") pod \"1a4df5d0-e6f7-403b-898a-46d3240f9100\" (UID: \"1a4df5d0-e6f7-403b-898a-46d3240f9100\") " Dec 03 16:27:05 crc kubenswrapper[4998]: I1203 16:27:05.913929 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a4df5d0-e6f7-403b-898a-46d3240f9100-logs" (OuterVolumeSpecName: "logs") pod "1a4df5d0-e6f7-403b-898a-46d3240f9100" (UID: "1a4df5d0-e6f7-403b-898a-46d3240f9100"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:27:05 crc kubenswrapper[4998]: I1203 16:27:05.922979 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a4df5d0-e6f7-403b-898a-46d3240f9100-kube-api-access-blm22" (OuterVolumeSpecName: "kube-api-access-blm22") pod "1a4df5d0-e6f7-403b-898a-46d3240f9100" (UID: "1a4df5d0-e6f7-403b-898a-46d3240f9100"). InnerVolumeSpecName "kube-api-access-blm22". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:27:05 crc kubenswrapper[4998]: I1203 16:27:05.948120 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a4df5d0-e6f7-403b-898a-46d3240f9100-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a4df5d0-e6f7-403b-898a-46d3240f9100" (UID: "1a4df5d0-e6f7-403b-898a-46d3240f9100"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:05 crc kubenswrapper[4998]: I1203 16:27:05.951091 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a4df5d0-e6f7-403b-898a-46d3240f9100-config-data" (OuterVolumeSpecName: "config-data") pod "1a4df5d0-e6f7-403b-898a-46d3240f9100" (UID: "1a4df5d0-e6f7-403b-898a-46d3240f9100"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.015794 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a4df5d0-e6f7-403b-898a-46d3240f9100-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.015831 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blm22\" (UniqueName: \"kubernetes.io/projected/1a4df5d0-e6f7-403b-898a-46d3240f9100-kube-api-access-blm22\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.015845 4998 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a4df5d0-e6f7-403b-898a-46d3240f9100-logs\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.015857 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a4df5d0-e6f7-403b-898a-46d3240f9100-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.699445 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.731424 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.742922 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.752515 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 16:27:06 crc kubenswrapper[4998]: E1203 16:27:06.752909 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a4df5d0-e6f7-403b-898a-46d3240f9100" containerName="nova-metadata-log" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.752925 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a4df5d0-e6f7-403b-898a-46d3240f9100" containerName="nova-metadata-log" Dec 03 16:27:06 crc kubenswrapper[4998]: E1203 16:27:06.752939 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a4df5d0-e6f7-403b-898a-46d3240f9100" containerName="nova-metadata-metadata" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.752946 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a4df5d0-e6f7-403b-898a-46d3240f9100" containerName="nova-metadata-metadata" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.753125 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a4df5d0-e6f7-403b-898a-46d3240f9100" containerName="nova-metadata-metadata" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.753150 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a4df5d0-e6f7-403b-898a-46d3240f9100" containerName="nova-metadata-log" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.754149 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.756659 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.757221 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.762601 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.833268 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab6acf83-6345-4166-8687-9cc00e958307-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ab6acf83-6345-4166-8687-9cc00e958307\") " pod="openstack/nova-metadata-0" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.833426 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab6acf83-6345-4166-8687-9cc00e958307-logs\") pod \"nova-metadata-0\" (UID: \"ab6acf83-6345-4166-8687-9cc00e958307\") " pod="openstack/nova-metadata-0" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.833590 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqxcj\" (UniqueName: \"kubernetes.io/projected/ab6acf83-6345-4166-8687-9cc00e958307-kube-api-access-vqxcj\") pod \"nova-metadata-0\" (UID: \"ab6acf83-6345-4166-8687-9cc00e958307\") " pod="openstack/nova-metadata-0" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.833793 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab6acf83-6345-4166-8687-9cc00e958307-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ab6acf83-6345-4166-8687-9cc00e958307\") " pod="openstack/nova-metadata-0" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.833882 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab6acf83-6345-4166-8687-9cc00e958307-config-data\") pod \"nova-metadata-0\" (UID: \"ab6acf83-6345-4166-8687-9cc00e958307\") " pod="openstack/nova-metadata-0" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.935378 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab6acf83-6345-4166-8687-9cc00e958307-config-data\") pod \"nova-metadata-0\" (UID: \"ab6acf83-6345-4166-8687-9cc00e958307\") " pod="openstack/nova-metadata-0" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.935446 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab6acf83-6345-4166-8687-9cc00e958307-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ab6acf83-6345-4166-8687-9cc00e958307\") " pod="openstack/nova-metadata-0" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.936281 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab6acf83-6345-4166-8687-9cc00e958307-logs\") pod \"nova-metadata-0\" (UID: \"ab6acf83-6345-4166-8687-9cc00e958307\") " pod="openstack/nova-metadata-0" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.936349 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqxcj\" (UniqueName: \"kubernetes.io/projected/ab6acf83-6345-4166-8687-9cc00e958307-kube-api-access-vqxcj\") pod \"nova-metadata-0\" (UID: \"ab6acf83-6345-4166-8687-9cc00e958307\") " pod="openstack/nova-metadata-0" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.936418 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab6acf83-6345-4166-8687-9cc00e958307-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ab6acf83-6345-4166-8687-9cc00e958307\") " pod="openstack/nova-metadata-0" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.936741 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab6acf83-6345-4166-8687-9cc00e958307-logs\") pod \"nova-metadata-0\" (UID: \"ab6acf83-6345-4166-8687-9cc00e958307\") " pod="openstack/nova-metadata-0" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.940711 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab6acf83-6345-4166-8687-9cc00e958307-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ab6acf83-6345-4166-8687-9cc00e958307\") " pod="openstack/nova-metadata-0" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.941427 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab6acf83-6345-4166-8687-9cc00e958307-config-data\") pod \"nova-metadata-0\" (UID: \"ab6acf83-6345-4166-8687-9cc00e958307\") " pod="openstack/nova-metadata-0" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.954430 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab6acf83-6345-4166-8687-9cc00e958307-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ab6acf83-6345-4166-8687-9cc00e958307\") " pod="openstack/nova-metadata-0" Dec 03 16:27:06 crc kubenswrapper[4998]: I1203 16:27:06.955916 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqxcj\" (UniqueName: \"kubernetes.io/projected/ab6acf83-6345-4166-8687-9cc00e958307-kube-api-access-vqxcj\") pod \"nova-metadata-0\" (UID: \"ab6acf83-6345-4166-8687-9cc00e958307\") " pod="openstack/nova-metadata-0" Dec 03 16:27:07 crc kubenswrapper[4998]: I1203 16:27:07.145694 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 16:27:07 crc kubenswrapper[4998]: W1203 16:27:07.674420 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab6acf83_6345_4166_8687_9cc00e958307.slice/crio-1332496c03680497139f6f7f6d9f2b0cd8209b6470f0e8dfe6fd6d7d79ad79d6 WatchSource:0}: Error finding container 1332496c03680497139f6f7f6d9f2b0cd8209b6470f0e8dfe6fd6d7d79ad79d6: Status 404 returned error can't find the container with id 1332496c03680497139f6f7f6d9f2b0cd8209b6470f0e8dfe6fd6d7d79ad79d6 Dec 03 16:27:07 crc kubenswrapper[4998]: I1203 16:27:07.689188 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a4df5d0-e6f7-403b-898a-46d3240f9100" path="/var/lib/kubelet/pods/1a4df5d0-e6f7-403b-898a-46d3240f9100/volumes" Dec 03 16:27:07 crc kubenswrapper[4998]: I1203 16:27:07.690289 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 16:27:07 crc kubenswrapper[4998]: I1203 16:27:07.715839 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ab6acf83-6345-4166-8687-9cc00e958307","Type":"ContainerStarted","Data":"1332496c03680497139f6f7f6d9f2b0cd8209b6470f0e8dfe6fd6d7d79ad79d6"} Dec 03 16:27:08 crc kubenswrapper[4998]: I1203 16:27:08.726990 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ab6acf83-6345-4166-8687-9cc00e958307","Type":"ContainerStarted","Data":"f0ecd8ae61ad219b2bccded8dbde99d8f98e8e1c8cd94b4df44b802259e2102c"} Dec 03 16:27:08 crc kubenswrapper[4998]: I1203 16:27:08.728012 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ab6acf83-6345-4166-8687-9cc00e958307","Type":"ContainerStarted","Data":"178848baf1d4fd0c8a090928e34ac8883063c414d0e2eb8f1c5221837b035f00"} Dec 03 16:27:08 crc kubenswrapper[4998]: I1203 16:27:08.751103 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.751085237 podStartE2EDuration="2.751085237s" podCreationTimestamp="2025-12-03 16:27:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:27:08.747424683 +0000 UTC m=+1407.359124906" watchObservedRunningTime="2025-12-03 16:27:08.751085237 +0000 UTC m=+1407.362785460" Dec 03 16:27:09 crc kubenswrapper[4998]: I1203 16:27:09.882176 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 16:27:09 crc kubenswrapper[4998]: I1203 16:27:09.882534 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 16:27:09 crc kubenswrapper[4998]: I1203 16:27:09.914020 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 16:27:09 crc kubenswrapper[4998]: I1203 16:27:09.958153 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 16:27:10 crc kubenswrapper[4998]: I1203 16:27:10.399943 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7995555d47-42cp9" Dec 03 16:27:10 crc kubenswrapper[4998]: I1203 16:27:10.470351 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85c94b455f-r5qz6"] Dec 03 16:27:10 crc kubenswrapper[4998]: I1203 16:27:10.470911 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" podUID="e22aa31f-72c8-45c5-bcbc-91c5ada645a3" containerName="dnsmasq-dns" containerID="cri-o://2af0ef43191ec1a4c35ffb9f96b92eb4a1259eaa63c5a31a2d279444fb0c33c6" gracePeriod=10 Dec 03 16:27:10 crc kubenswrapper[4998]: I1203 16:27:10.758374 4998 generic.go:334] "Generic (PLEG): container finished" podID="0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d" containerID="e9f429e5e1f1bd89d2bb9cca65fdc0eebb511a8d993f432a17e22b665be35a4c" exitCode=0 Dec 03 16:27:10 crc kubenswrapper[4998]: I1203 16:27:10.758472 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-jzsvl" event={"ID":"0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d","Type":"ContainerDied","Data":"e9f429e5e1f1bd89d2bb9cca65fdc0eebb511a8d993f432a17e22b665be35a4c"} Dec 03 16:27:10 crc kubenswrapper[4998]: I1203 16:27:10.793545 4998 generic.go:334] "Generic (PLEG): container finished" podID="e22aa31f-72c8-45c5-bcbc-91c5ada645a3" containerID="2af0ef43191ec1a4c35ffb9f96b92eb4a1259eaa63c5a31a2d279444fb0c33c6" exitCode=0 Dec 03 16:27:10 crc kubenswrapper[4998]: I1203 16:27:10.793597 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" event={"ID":"e22aa31f-72c8-45c5-bcbc-91c5ada645a3","Type":"ContainerDied","Data":"2af0ef43191ec1a4c35ffb9f96b92eb4a1259eaa63c5a31a2d279444fb0c33c6"} Dec 03 16:27:10 crc kubenswrapper[4998]: I1203 16:27:10.830312 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 16:27:10 crc kubenswrapper[4998]: I1203 16:27:10.966193 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" Dec 03 16:27:10 crc kubenswrapper[4998]: I1203 16:27:10.978975 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f9c25919-a65d-4e03-93e1-4d19a42b3d28" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.207:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 16:27:10 crc kubenswrapper[4998]: I1203 16:27:10.979042 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f9c25919-a65d-4e03-93e1-4d19a42b3d28" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.207:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 16:27:11 crc kubenswrapper[4998]: I1203 16:27:11.123319 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-config\") pod \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\" (UID: \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\") " Dec 03 16:27:11 crc kubenswrapper[4998]: I1203 16:27:11.123382 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfw27\" (UniqueName: \"kubernetes.io/projected/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-kube-api-access-tfw27\") pod \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\" (UID: \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\") " Dec 03 16:27:11 crc kubenswrapper[4998]: I1203 16:27:11.123431 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-ovsdbserver-nb\") pod \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\" (UID: \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\") " Dec 03 16:27:11 crc kubenswrapper[4998]: I1203 16:27:11.123449 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-ovsdbserver-sb\") pod \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\" (UID: \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\") " Dec 03 16:27:11 crc kubenswrapper[4998]: I1203 16:27:11.123539 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-dns-svc\") pod \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\" (UID: \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\") " Dec 03 16:27:11 crc kubenswrapper[4998]: I1203 16:27:11.123632 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-dns-swift-storage-0\") pod \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\" (UID: \"e22aa31f-72c8-45c5-bcbc-91c5ada645a3\") " Dec 03 16:27:11 crc kubenswrapper[4998]: I1203 16:27:11.148480 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-kube-api-access-tfw27" (OuterVolumeSpecName: "kube-api-access-tfw27") pod "e22aa31f-72c8-45c5-bcbc-91c5ada645a3" (UID: "e22aa31f-72c8-45c5-bcbc-91c5ada645a3"). InnerVolumeSpecName "kube-api-access-tfw27". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:27:11 crc kubenswrapper[4998]: I1203 16:27:11.200664 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e22aa31f-72c8-45c5-bcbc-91c5ada645a3" (UID: "e22aa31f-72c8-45c5-bcbc-91c5ada645a3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:27:11 crc kubenswrapper[4998]: I1203 16:27:11.210343 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e22aa31f-72c8-45c5-bcbc-91c5ada645a3" (UID: "e22aa31f-72c8-45c5-bcbc-91c5ada645a3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:27:11 crc kubenswrapper[4998]: I1203 16:27:11.217535 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e22aa31f-72c8-45c5-bcbc-91c5ada645a3" (UID: "e22aa31f-72c8-45c5-bcbc-91c5ada645a3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:27:11 crc kubenswrapper[4998]: I1203 16:27:11.226911 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e22aa31f-72c8-45c5-bcbc-91c5ada645a3" (UID: "e22aa31f-72c8-45c5-bcbc-91c5ada645a3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:27:11 crc kubenswrapper[4998]: I1203 16:27:11.227865 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfw27\" (UniqueName: \"kubernetes.io/projected/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-kube-api-access-tfw27\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:11 crc kubenswrapper[4998]: I1203 16:27:11.227893 4998 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:11 crc kubenswrapper[4998]: I1203 16:27:11.227909 4998 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:11 crc kubenswrapper[4998]: I1203 16:27:11.227920 4998 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:11 crc kubenswrapper[4998]: I1203 16:27:11.227932 4998 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:11 crc kubenswrapper[4998]: I1203 16:27:11.231196 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-config" (OuterVolumeSpecName: "config") pod "e22aa31f-72c8-45c5-bcbc-91c5ada645a3" (UID: "e22aa31f-72c8-45c5-bcbc-91c5ada645a3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:27:11 crc kubenswrapper[4998]: I1203 16:27:11.329866 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e22aa31f-72c8-45c5-bcbc-91c5ada645a3-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:11 crc kubenswrapper[4998]: I1203 16:27:11.808458 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" Dec 03 16:27:11 crc kubenswrapper[4998]: I1203 16:27:11.808992 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" event={"ID":"e22aa31f-72c8-45c5-bcbc-91c5ada645a3","Type":"ContainerDied","Data":"4e293e6dfdf06990c533afb15796569a0e9fe89f5ee647a6bc6041e92a23f7ec"} Dec 03 16:27:11 crc kubenswrapper[4998]: I1203 16:27:11.809062 4998 scope.go:117] "RemoveContainer" containerID="2af0ef43191ec1a4c35ffb9f96b92eb4a1259eaa63c5a31a2d279444fb0c33c6" Dec 03 16:27:11 crc kubenswrapper[4998]: I1203 16:27:11.842863 4998 scope.go:117] "RemoveContainer" containerID="7d7328b400c9283c54c6fc3ae67e3d0c99c825f5e02471ab881dabacee8513d5" Dec 03 16:27:11 crc kubenswrapper[4998]: I1203 16:27:11.842989 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85c94b455f-r5qz6"] Dec 03 16:27:11 crc kubenswrapper[4998]: I1203 16:27:11.877444 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85c94b455f-r5qz6"] Dec 03 16:27:12 crc kubenswrapper[4998]: I1203 16:27:12.147231 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 16:27:12 crc kubenswrapper[4998]: I1203 16:27:12.147543 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 16:27:12 crc kubenswrapper[4998]: I1203 16:27:12.181519 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-jzsvl" Dec 03 16:27:12 crc kubenswrapper[4998]: I1203 16:27:12.347356 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d-combined-ca-bundle\") pod \"0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d\" (UID: \"0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d\") " Dec 03 16:27:12 crc kubenswrapper[4998]: I1203 16:27:12.347470 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7z7ls\" (UniqueName: \"kubernetes.io/projected/0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d-kube-api-access-7z7ls\") pod \"0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d\" (UID: \"0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d\") " Dec 03 16:27:12 crc kubenswrapper[4998]: I1203 16:27:12.347502 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d-config-data\") pod \"0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d\" (UID: \"0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d\") " Dec 03 16:27:12 crc kubenswrapper[4998]: I1203 16:27:12.347559 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d-scripts\") pod \"0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d\" (UID: \"0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d\") " Dec 03 16:27:12 crc kubenswrapper[4998]: I1203 16:27:12.351905 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d-scripts" (OuterVolumeSpecName: "scripts") pod "0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d" (UID: "0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:12 crc kubenswrapper[4998]: I1203 16:27:12.351915 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d-kube-api-access-7z7ls" (OuterVolumeSpecName: "kube-api-access-7z7ls") pod "0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d" (UID: "0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d"). InnerVolumeSpecName "kube-api-access-7z7ls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:27:12 crc kubenswrapper[4998]: I1203 16:27:12.383848 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d-config-data" (OuterVolumeSpecName: "config-data") pod "0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d" (UID: "0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:12 crc kubenswrapper[4998]: I1203 16:27:12.399598 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d" (UID: "0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:12 crc kubenswrapper[4998]: I1203 16:27:12.449224 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:12 crc kubenswrapper[4998]: I1203 16:27:12.449254 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7z7ls\" (UniqueName: \"kubernetes.io/projected/0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d-kube-api-access-7z7ls\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:12 crc kubenswrapper[4998]: I1203 16:27:12.449264 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:12 crc kubenswrapper[4998]: I1203 16:27:12.449273 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:12 crc kubenswrapper[4998]: I1203 16:27:12.829295 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-jzsvl" event={"ID":"0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d","Type":"ContainerDied","Data":"954a0b7734993941d2aefbb239e007d250add17e72345ca486ca45fd68b68e0a"} Dec 03 16:27:12 crc kubenswrapper[4998]: I1203 16:27:12.829343 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="954a0b7734993941d2aefbb239e007d250add17e72345ca486ca45fd68b68e0a" Dec 03 16:27:12 crc kubenswrapper[4998]: I1203 16:27:12.829408 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-jzsvl" Dec 03 16:27:12 crc kubenswrapper[4998]: I1203 16:27:12.970392 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 16:27:12 crc kubenswrapper[4998]: I1203 16:27:12.970716 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f9c25919-a65d-4e03-93e1-4d19a42b3d28" containerName="nova-api-log" containerID="cri-o://458902618fd0712551d5f6d404870bed957eecaa53187f738e42909f6f22cdd4" gracePeriod=30 Dec 03 16:27:12 crc kubenswrapper[4998]: I1203 16:27:12.971507 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f9c25919-a65d-4e03-93e1-4d19a42b3d28" containerName="nova-api-api" containerID="cri-o://c0df6738509fa9424f3e8f55f7f3a2cbfe198c9d93aa4df0fdf48d5497eea9f7" gracePeriod=30 Dec 03 16:27:13 crc kubenswrapper[4998]: I1203 16:27:13.006140 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 16:27:13 crc kubenswrapper[4998]: I1203 16:27:13.006532 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="37109c33-9436-40d5-aa8f-bb7e7c3a0f83" containerName="nova-scheduler-scheduler" containerID="cri-o://682763cd65cf4804dc534645f561db0b3daf020b18f1b16c9b083082fae8e47d" gracePeriod=30 Dec 03 16:27:13 crc kubenswrapper[4998]: I1203 16:27:13.036321 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 16:27:13 crc kubenswrapper[4998]: I1203 16:27:13.036537 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ab6acf83-6345-4166-8687-9cc00e958307" containerName="nova-metadata-log" containerID="cri-o://178848baf1d4fd0c8a090928e34ac8883063c414d0e2eb8f1c5221837b035f00" gracePeriod=30 Dec 03 16:27:13 crc kubenswrapper[4998]: I1203 16:27:13.036632 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ab6acf83-6345-4166-8687-9cc00e958307" containerName="nova-metadata-metadata" containerID="cri-o://f0ecd8ae61ad219b2bccded8dbde99d8f98e8e1c8cd94b4df44b802259e2102c" gracePeriod=30 Dec 03 16:27:13 crc kubenswrapper[4998]: I1203 16:27:13.712452 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e22aa31f-72c8-45c5-bcbc-91c5ada645a3" path="/var/lib/kubelet/pods/e22aa31f-72c8-45c5-bcbc-91c5ada645a3/volumes" Dec 03 16:27:13 crc kubenswrapper[4998]: I1203 16:27:13.851910 4998 generic.go:334] "Generic (PLEG): container finished" podID="ab6acf83-6345-4166-8687-9cc00e958307" containerID="f0ecd8ae61ad219b2bccded8dbde99d8f98e8e1c8cd94b4df44b802259e2102c" exitCode=0 Dec 03 16:27:13 crc kubenswrapper[4998]: I1203 16:27:13.851974 4998 generic.go:334] "Generic (PLEG): container finished" podID="ab6acf83-6345-4166-8687-9cc00e958307" containerID="178848baf1d4fd0c8a090928e34ac8883063c414d0e2eb8f1c5221837b035f00" exitCode=143 Dec 03 16:27:13 crc kubenswrapper[4998]: I1203 16:27:13.852045 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ab6acf83-6345-4166-8687-9cc00e958307","Type":"ContainerDied","Data":"f0ecd8ae61ad219b2bccded8dbde99d8f98e8e1c8cd94b4df44b802259e2102c"} Dec 03 16:27:13 crc kubenswrapper[4998]: I1203 16:27:13.852109 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ab6acf83-6345-4166-8687-9cc00e958307","Type":"ContainerDied","Data":"178848baf1d4fd0c8a090928e34ac8883063c414d0e2eb8f1c5221837b035f00"} Dec 03 16:27:13 crc kubenswrapper[4998]: I1203 16:27:13.854784 4998 generic.go:334] "Generic (PLEG): container finished" podID="f9c25919-a65d-4e03-93e1-4d19a42b3d28" containerID="458902618fd0712551d5f6d404870bed957eecaa53187f738e42909f6f22cdd4" exitCode=143 Dec 03 16:27:13 crc kubenswrapper[4998]: I1203 16:27:13.854825 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f9c25919-a65d-4e03-93e1-4d19a42b3d28","Type":"ContainerDied","Data":"458902618fd0712551d5f6d404870bed957eecaa53187f738e42909f6f22cdd4"} Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.132729 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.286903 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab6acf83-6345-4166-8687-9cc00e958307-nova-metadata-tls-certs\") pod \"ab6acf83-6345-4166-8687-9cc00e958307\" (UID: \"ab6acf83-6345-4166-8687-9cc00e958307\") " Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.286943 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab6acf83-6345-4166-8687-9cc00e958307-logs\") pod \"ab6acf83-6345-4166-8687-9cc00e958307\" (UID: \"ab6acf83-6345-4166-8687-9cc00e958307\") " Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.287039 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqxcj\" (UniqueName: \"kubernetes.io/projected/ab6acf83-6345-4166-8687-9cc00e958307-kube-api-access-vqxcj\") pod \"ab6acf83-6345-4166-8687-9cc00e958307\" (UID: \"ab6acf83-6345-4166-8687-9cc00e958307\") " Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.287068 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab6acf83-6345-4166-8687-9cc00e958307-config-data\") pod \"ab6acf83-6345-4166-8687-9cc00e958307\" (UID: \"ab6acf83-6345-4166-8687-9cc00e958307\") " Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.287191 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab6acf83-6345-4166-8687-9cc00e958307-combined-ca-bundle\") pod \"ab6acf83-6345-4166-8687-9cc00e958307\" (UID: \"ab6acf83-6345-4166-8687-9cc00e958307\") " Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.287729 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab6acf83-6345-4166-8687-9cc00e958307-logs" (OuterVolumeSpecName: "logs") pod "ab6acf83-6345-4166-8687-9cc00e958307" (UID: "ab6acf83-6345-4166-8687-9cc00e958307"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.294142 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab6acf83-6345-4166-8687-9cc00e958307-kube-api-access-vqxcj" (OuterVolumeSpecName: "kube-api-access-vqxcj") pod "ab6acf83-6345-4166-8687-9cc00e958307" (UID: "ab6acf83-6345-4166-8687-9cc00e958307"). InnerVolumeSpecName "kube-api-access-vqxcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.324020 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab6acf83-6345-4166-8687-9cc00e958307-config-data" (OuterVolumeSpecName: "config-data") pod "ab6acf83-6345-4166-8687-9cc00e958307" (UID: "ab6acf83-6345-4166-8687-9cc00e958307"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.329717 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab6acf83-6345-4166-8687-9cc00e958307-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab6acf83-6345-4166-8687-9cc00e958307" (UID: "ab6acf83-6345-4166-8687-9cc00e958307"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.352158 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab6acf83-6345-4166-8687-9cc00e958307-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "ab6acf83-6345-4166-8687-9cc00e958307" (UID: "ab6acf83-6345-4166-8687-9cc00e958307"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.389323 4998 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab6acf83-6345-4166-8687-9cc00e958307-logs\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.389352 4998 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab6acf83-6345-4166-8687-9cc00e958307-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.389363 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqxcj\" (UniqueName: \"kubernetes.io/projected/ab6acf83-6345-4166-8687-9cc00e958307-kube-api-access-vqxcj\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.389373 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab6acf83-6345-4166-8687-9cc00e958307-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.389383 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab6acf83-6345-4166-8687-9cc00e958307-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.868170 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ab6acf83-6345-4166-8687-9cc00e958307","Type":"ContainerDied","Data":"1332496c03680497139f6f7f6d9f2b0cd8209b6470f0e8dfe6fd6d7d79ad79d6"} Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.868237 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.868655 4998 scope.go:117] "RemoveContainer" containerID="f0ecd8ae61ad219b2bccded8dbde99d8f98e8e1c8cd94b4df44b802259e2102c" Dec 03 16:27:14 crc kubenswrapper[4998]: E1203 16:27:14.910998 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="682763cd65cf4804dc534645f561db0b3daf020b18f1b16c9b083082fae8e47d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.912524 4998 scope.go:117] "RemoveContainer" containerID="178848baf1d4fd0c8a090928e34ac8883063c414d0e2eb8f1c5221837b035f00" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.913730 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 16:27:14 crc kubenswrapper[4998]: E1203 16:27:14.913743 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="682763cd65cf4804dc534645f561db0b3daf020b18f1b16c9b083082fae8e47d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 16:27:14 crc kubenswrapper[4998]: E1203 16:27:14.918949 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="682763cd65cf4804dc534645f561db0b3daf020b18f1b16c9b083082fae8e47d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 16:27:14 crc kubenswrapper[4998]: E1203 16:27:14.919044 4998 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="37109c33-9436-40d5-aa8f-bb7e7c3a0f83" containerName="nova-scheduler-scheduler" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.940828 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.955156 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 16:27:14 crc kubenswrapper[4998]: E1203 16:27:14.955548 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d" containerName="nova-manage" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.955564 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d" containerName="nova-manage" Dec 03 16:27:14 crc kubenswrapper[4998]: E1203 16:27:14.955585 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab6acf83-6345-4166-8687-9cc00e958307" containerName="nova-metadata-metadata" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.955591 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab6acf83-6345-4166-8687-9cc00e958307" containerName="nova-metadata-metadata" Dec 03 16:27:14 crc kubenswrapper[4998]: E1203 16:27:14.955600 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e22aa31f-72c8-45c5-bcbc-91c5ada645a3" containerName="init" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.955606 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="e22aa31f-72c8-45c5-bcbc-91c5ada645a3" containerName="init" Dec 03 16:27:14 crc kubenswrapper[4998]: E1203 16:27:14.955622 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab6acf83-6345-4166-8687-9cc00e958307" containerName="nova-metadata-log" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.955628 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab6acf83-6345-4166-8687-9cc00e958307" containerName="nova-metadata-log" Dec 03 16:27:14 crc kubenswrapper[4998]: E1203 16:27:14.955643 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e22aa31f-72c8-45c5-bcbc-91c5ada645a3" containerName="dnsmasq-dns" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.955649 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="e22aa31f-72c8-45c5-bcbc-91c5ada645a3" containerName="dnsmasq-dns" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.956017 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab6acf83-6345-4166-8687-9cc00e958307" containerName="nova-metadata-log" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.956035 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab6acf83-6345-4166-8687-9cc00e958307" containerName="nova-metadata-metadata" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.956048 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d" containerName="nova-manage" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.956070 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="e22aa31f-72c8-45c5-bcbc-91c5ada645a3" containerName="dnsmasq-dns" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.957249 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.961362 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.969342 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 16:27:14 crc kubenswrapper[4998]: I1203 16:27:14.974914 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 16:27:15 crc kubenswrapper[4998]: I1203 16:27:15.104939 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd010875-398a-48fc-8fa5-569126c4f67f-logs\") pod \"nova-metadata-0\" (UID: \"cd010875-398a-48fc-8fa5-569126c4f67f\") " pod="openstack/nova-metadata-0" Dec 03 16:27:15 crc kubenswrapper[4998]: I1203 16:27:15.104996 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd010875-398a-48fc-8fa5-569126c4f67f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"cd010875-398a-48fc-8fa5-569126c4f67f\") " pod="openstack/nova-metadata-0" Dec 03 16:27:15 crc kubenswrapper[4998]: I1203 16:27:15.105349 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd010875-398a-48fc-8fa5-569126c4f67f-config-data\") pod \"nova-metadata-0\" (UID: \"cd010875-398a-48fc-8fa5-569126c4f67f\") " pod="openstack/nova-metadata-0" Dec 03 16:27:15 crc kubenswrapper[4998]: I1203 16:27:15.105448 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szwsj\" (UniqueName: \"kubernetes.io/projected/cd010875-398a-48fc-8fa5-569126c4f67f-kube-api-access-szwsj\") pod \"nova-metadata-0\" (UID: \"cd010875-398a-48fc-8fa5-569126c4f67f\") " pod="openstack/nova-metadata-0" Dec 03 16:27:15 crc kubenswrapper[4998]: I1203 16:27:15.105521 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd010875-398a-48fc-8fa5-569126c4f67f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cd010875-398a-48fc-8fa5-569126c4f67f\") " pod="openstack/nova-metadata-0" Dec 03 16:27:15 crc kubenswrapper[4998]: I1203 16:27:15.207519 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd010875-398a-48fc-8fa5-569126c4f67f-logs\") pod \"nova-metadata-0\" (UID: \"cd010875-398a-48fc-8fa5-569126c4f67f\") " pod="openstack/nova-metadata-0" Dec 03 16:27:15 crc kubenswrapper[4998]: I1203 16:27:15.207600 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd010875-398a-48fc-8fa5-569126c4f67f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"cd010875-398a-48fc-8fa5-569126c4f67f\") " pod="openstack/nova-metadata-0" Dec 03 16:27:15 crc kubenswrapper[4998]: I1203 16:27:15.207700 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd010875-398a-48fc-8fa5-569126c4f67f-config-data\") pod \"nova-metadata-0\" (UID: \"cd010875-398a-48fc-8fa5-569126c4f67f\") " pod="openstack/nova-metadata-0" Dec 03 16:27:15 crc kubenswrapper[4998]: I1203 16:27:15.207797 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szwsj\" (UniqueName: \"kubernetes.io/projected/cd010875-398a-48fc-8fa5-569126c4f67f-kube-api-access-szwsj\") pod \"nova-metadata-0\" (UID: \"cd010875-398a-48fc-8fa5-569126c4f67f\") " pod="openstack/nova-metadata-0" Dec 03 16:27:15 crc kubenswrapper[4998]: I1203 16:27:15.207861 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd010875-398a-48fc-8fa5-569126c4f67f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cd010875-398a-48fc-8fa5-569126c4f67f\") " pod="openstack/nova-metadata-0" Dec 03 16:27:15 crc kubenswrapper[4998]: I1203 16:27:15.211291 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd010875-398a-48fc-8fa5-569126c4f67f-logs\") pod \"nova-metadata-0\" (UID: \"cd010875-398a-48fc-8fa5-569126c4f67f\") " pod="openstack/nova-metadata-0" Dec 03 16:27:15 crc kubenswrapper[4998]: I1203 16:27:15.215447 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd010875-398a-48fc-8fa5-569126c4f67f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"cd010875-398a-48fc-8fa5-569126c4f67f\") " pod="openstack/nova-metadata-0" Dec 03 16:27:15 crc kubenswrapper[4998]: I1203 16:27:15.216932 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd010875-398a-48fc-8fa5-569126c4f67f-config-data\") pod \"nova-metadata-0\" (UID: \"cd010875-398a-48fc-8fa5-569126c4f67f\") " pod="openstack/nova-metadata-0" Dec 03 16:27:15 crc kubenswrapper[4998]: I1203 16:27:15.217395 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd010875-398a-48fc-8fa5-569126c4f67f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cd010875-398a-48fc-8fa5-569126c4f67f\") " pod="openstack/nova-metadata-0" Dec 03 16:27:15 crc kubenswrapper[4998]: I1203 16:27:15.245999 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szwsj\" (UniqueName: \"kubernetes.io/projected/cd010875-398a-48fc-8fa5-569126c4f67f-kube-api-access-szwsj\") pod \"nova-metadata-0\" (UID: \"cd010875-398a-48fc-8fa5-569126c4f67f\") " pod="openstack/nova-metadata-0" Dec 03 16:27:15 crc kubenswrapper[4998]: I1203 16:27:15.288532 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 16:27:15 crc kubenswrapper[4998]: I1203 16:27:15.692831 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab6acf83-6345-4166-8687-9cc00e958307" path="/var/lib/kubelet/pods/ab6acf83-6345-4166-8687-9cc00e958307/volumes" Dec 03 16:27:15 crc kubenswrapper[4998]: I1203 16:27:15.756038 4998 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-85c94b455f-r5qz6" podUID="e22aa31f-72c8-45c5-bcbc-91c5ada645a3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.180:5353: i/o timeout" Dec 03 16:27:15 crc kubenswrapper[4998]: I1203 16:27:15.796197 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 16:27:15 crc kubenswrapper[4998]: I1203 16:27:15.804720 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 16:27:15 crc kubenswrapper[4998]: I1203 16:27:15.998170 4998 generic.go:334] "Generic (PLEG): container finished" podID="f9c25919-a65d-4e03-93e1-4d19a42b3d28" containerID="c0df6738509fa9424f3e8f55f7f3a2cbfe198c9d93aa4df0fdf48d5497eea9f7" exitCode=0 Dec 03 16:27:15 crc kubenswrapper[4998]: I1203 16:27:15.998299 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f9c25919-a65d-4e03-93e1-4d19a42b3d28","Type":"ContainerDied","Data":"c0df6738509fa9424f3e8f55f7f3a2cbfe198c9d93aa4df0fdf48d5497eea9f7"} Dec 03 16:27:16 crc kubenswrapper[4998]: I1203 16:27:16.011553 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cd010875-398a-48fc-8fa5-569126c4f67f","Type":"ContainerStarted","Data":"bbc157a93f57519359c12ddddcafb73d83738537de2687889c41da4eee74a7e0"} Dec 03 16:27:16 crc kubenswrapper[4998]: I1203 16:27:16.421206 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 16:27:16 crc kubenswrapper[4998]: I1203 16:27:16.554528 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9c25919-a65d-4e03-93e1-4d19a42b3d28-combined-ca-bundle\") pod \"f9c25919-a65d-4e03-93e1-4d19a42b3d28\" (UID: \"f9c25919-a65d-4e03-93e1-4d19a42b3d28\") " Dec 03 16:27:16 crc kubenswrapper[4998]: I1203 16:27:16.554679 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9c25919-a65d-4e03-93e1-4d19a42b3d28-logs\") pod \"f9c25919-a65d-4e03-93e1-4d19a42b3d28\" (UID: \"f9c25919-a65d-4e03-93e1-4d19a42b3d28\") " Dec 03 16:27:16 crc kubenswrapper[4998]: I1203 16:27:16.554703 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8grrm\" (UniqueName: \"kubernetes.io/projected/f9c25919-a65d-4e03-93e1-4d19a42b3d28-kube-api-access-8grrm\") pod \"f9c25919-a65d-4e03-93e1-4d19a42b3d28\" (UID: \"f9c25919-a65d-4e03-93e1-4d19a42b3d28\") " Dec 03 16:27:16 crc kubenswrapper[4998]: I1203 16:27:16.554793 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9c25919-a65d-4e03-93e1-4d19a42b3d28-config-data\") pod \"f9c25919-a65d-4e03-93e1-4d19a42b3d28\" (UID: \"f9c25919-a65d-4e03-93e1-4d19a42b3d28\") " Dec 03 16:27:16 crc kubenswrapper[4998]: I1203 16:27:16.555416 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9c25919-a65d-4e03-93e1-4d19a42b3d28-logs" (OuterVolumeSpecName: "logs") pod "f9c25919-a65d-4e03-93e1-4d19a42b3d28" (UID: "f9c25919-a65d-4e03-93e1-4d19a42b3d28"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:27:16 crc kubenswrapper[4998]: I1203 16:27:16.555837 4998 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f9c25919-a65d-4e03-93e1-4d19a42b3d28-logs\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:16 crc kubenswrapper[4998]: I1203 16:27:16.559651 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9c25919-a65d-4e03-93e1-4d19a42b3d28-kube-api-access-8grrm" (OuterVolumeSpecName: "kube-api-access-8grrm") pod "f9c25919-a65d-4e03-93e1-4d19a42b3d28" (UID: "f9c25919-a65d-4e03-93e1-4d19a42b3d28"). InnerVolumeSpecName "kube-api-access-8grrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:27:16 crc kubenswrapper[4998]: I1203 16:27:16.596905 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9c25919-a65d-4e03-93e1-4d19a42b3d28-config-data" (OuterVolumeSpecName: "config-data") pod "f9c25919-a65d-4e03-93e1-4d19a42b3d28" (UID: "f9c25919-a65d-4e03-93e1-4d19a42b3d28"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:16 crc kubenswrapper[4998]: I1203 16:27:16.622901 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9c25919-a65d-4e03-93e1-4d19a42b3d28-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f9c25919-a65d-4e03-93e1-4d19a42b3d28" (UID: "f9c25919-a65d-4e03-93e1-4d19a42b3d28"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:16 crc kubenswrapper[4998]: I1203 16:27:16.657184 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9c25919-a65d-4e03-93e1-4d19a42b3d28-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:16 crc kubenswrapper[4998]: I1203 16:27:16.657251 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8grrm\" (UniqueName: \"kubernetes.io/projected/f9c25919-a65d-4e03-93e1-4d19a42b3d28-kube-api-access-8grrm\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:16 crc kubenswrapper[4998]: I1203 16:27:16.657263 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9c25919-a65d-4e03-93e1-4d19a42b3d28-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.021780 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cd010875-398a-48fc-8fa5-569126c4f67f","Type":"ContainerStarted","Data":"b276c4c9c27b1341e38931073a93f1259bd29c8f45fc78cfe649b567c556b32c"} Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.022064 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cd010875-398a-48fc-8fa5-569126c4f67f","Type":"ContainerStarted","Data":"3598305bb163d8ec273d76c345836c5e48807e343bbabc183e9bb5c7b73b3df8"} Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.024936 4998 generic.go:334] "Generic (PLEG): container finished" podID="86d54b38-8913-4d18-92ee-490aceef1677" containerID="ae7a2f32fd866e6275ac3717266de1857c58450ea3f2547815e726e96f085b84" exitCode=0 Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.024985 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vvjng" event={"ID":"86d54b38-8913-4d18-92ee-490aceef1677","Type":"ContainerDied","Data":"ae7a2f32fd866e6275ac3717266de1857c58450ea3f2547815e726e96f085b84"} Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.026767 4998 generic.go:334] "Generic (PLEG): container finished" podID="37109c33-9436-40d5-aa8f-bb7e7c3a0f83" containerID="682763cd65cf4804dc534645f561db0b3daf020b18f1b16c9b083082fae8e47d" exitCode=0 Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.026807 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"37109c33-9436-40d5-aa8f-bb7e7c3a0f83","Type":"ContainerDied","Data":"682763cd65cf4804dc534645f561db0b3daf020b18f1b16c9b083082fae8e47d"} Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.029485 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f9c25919-a65d-4e03-93e1-4d19a42b3d28","Type":"ContainerDied","Data":"7e2d92aa5903e257688bb1e797558969d4984850a2aad26dd9dc787ba9bc89d7"} Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.029517 4998 scope.go:117] "RemoveContainer" containerID="c0df6738509fa9424f3e8f55f7f3a2cbfe198c9d93aa4df0fdf48d5497eea9f7" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.029639 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.049905 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.04988195 podStartE2EDuration="3.04988195s" podCreationTimestamp="2025-12-03 16:27:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:27:17.039601349 +0000 UTC m=+1415.651301572" watchObservedRunningTime="2025-12-03 16:27:17.04988195 +0000 UTC m=+1415.661582173" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.066259 4998 scope.go:117] "RemoveContainer" containerID="458902618fd0712551d5f6d404870bed957eecaa53187f738e42909f6f22cdd4" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.098208 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.118408 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.128247 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 16:27:17 crc kubenswrapper[4998]: E1203 16:27:17.128821 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9c25919-a65d-4e03-93e1-4d19a42b3d28" containerName="nova-api-log" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.128842 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9c25919-a65d-4e03-93e1-4d19a42b3d28" containerName="nova-api-log" Dec 03 16:27:17 crc kubenswrapper[4998]: E1203 16:27:17.128863 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9c25919-a65d-4e03-93e1-4d19a42b3d28" containerName="nova-api-api" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.128872 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9c25919-a65d-4e03-93e1-4d19a42b3d28" containerName="nova-api-api" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.136109 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9c25919-a65d-4e03-93e1-4d19a42b3d28" containerName="nova-api-log" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.136155 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9c25919-a65d-4e03-93e1-4d19a42b3d28" containerName="nova-api-api" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.137692 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.142220 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.144529 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.270370 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5\") " pod="openstack/nova-api-0" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.270568 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72qtb\" (UniqueName: \"kubernetes.io/projected/fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5-kube-api-access-72qtb\") pod \"nova-api-0\" (UID: \"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5\") " pod="openstack/nova-api-0" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.270627 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5-logs\") pod \"nova-api-0\" (UID: \"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5\") " pod="openstack/nova-api-0" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.270813 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5-config-data\") pod \"nova-api-0\" (UID: \"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5\") " pod="openstack/nova-api-0" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.372830 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5-config-data\") pod \"nova-api-0\" (UID: \"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5\") " pod="openstack/nova-api-0" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.372900 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5\") " pod="openstack/nova-api-0" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.373001 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72qtb\" (UniqueName: \"kubernetes.io/projected/fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5-kube-api-access-72qtb\") pod \"nova-api-0\" (UID: \"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5\") " pod="openstack/nova-api-0" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.373033 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5-logs\") pod \"nova-api-0\" (UID: \"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5\") " pod="openstack/nova-api-0" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.373410 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5-logs\") pod \"nova-api-0\" (UID: \"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5\") " pod="openstack/nova-api-0" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.381480 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5\") " pod="openstack/nova-api-0" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.381615 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5-config-data\") pod \"nova-api-0\" (UID: \"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5\") " pod="openstack/nova-api-0" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.390271 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72qtb\" (UniqueName: \"kubernetes.io/projected/fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5-kube-api-access-72qtb\") pod \"nova-api-0\" (UID: \"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5\") " pod="openstack/nova-api-0" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.457216 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.575272 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.677600 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lglp\" (UniqueName: \"kubernetes.io/projected/37109c33-9436-40d5-aa8f-bb7e7c3a0f83-kube-api-access-8lglp\") pod \"37109c33-9436-40d5-aa8f-bb7e7c3a0f83\" (UID: \"37109c33-9436-40d5-aa8f-bb7e7c3a0f83\") " Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.677742 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37109c33-9436-40d5-aa8f-bb7e7c3a0f83-config-data\") pod \"37109c33-9436-40d5-aa8f-bb7e7c3a0f83\" (UID: \"37109c33-9436-40d5-aa8f-bb7e7c3a0f83\") " Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.677924 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37109c33-9436-40d5-aa8f-bb7e7c3a0f83-combined-ca-bundle\") pod \"37109c33-9436-40d5-aa8f-bb7e7c3a0f83\" (UID: \"37109c33-9436-40d5-aa8f-bb7e7c3a0f83\") " Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.685912 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37109c33-9436-40d5-aa8f-bb7e7c3a0f83-kube-api-access-8lglp" (OuterVolumeSpecName: "kube-api-access-8lglp") pod "37109c33-9436-40d5-aa8f-bb7e7c3a0f83" (UID: "37109c33-9436-40d5-aa8f-bb7e7c3a0f83"). InnerVolumeSpecName "kube-api-access-8lglp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.691234 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9c25919-a65d-4e03-93e1-4d19a42b3d28" path="/var/lib/kubelet/pods/f9c25919-a65d-4e03-93e1-4d19a42b3d28/volumes" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.715534 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37109c33-9436-40d5-aa8f-bb7e7c3a0f83-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37109c33-9436-40d5-aa8f-bb7e7c3a0f83" (UID: "37109c33-9436-40d5-aa8f-bb7e7c3a0f83"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.715922 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37109c33-9436-40d5-aa8f-bb7e7c3a0f83-config-data" (OuterVolumeSpecName: "config-data") pod "37109c33-9436-40d5-aa8f-bb7e7c3a0f83" (UID: "37109c33-9436-40d5-aa8f-bb7e7c3a0f83"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.782008 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37109c33-9436-40d5-aa8f-bb7e7c3a0f83-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.782048 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lglp\" (UniqueName: \"kubernetes.io/projected/37109c33-9436-40d5-aa8f-bb7e7c3a0f83-kube-api-access-8lglp\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.782063 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37109c33-9436-40d5-aa8f-bb7e7c3a0f83-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:17 crc kubenswrapper[4998]: I1203 16:27:17.959988 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.041819 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5","Type":"ContainerStarted","Data":"fe2476b33790a54d9f75509e1d7bcaafabffc92c7cfc3a64f71fc8224a71dc65"} Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.044514 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"37109c33-9436-40d5-aa8f-bb7e7c3a0f83","Type":"ContainerDied","Data":"2bd0a3caf659a72d355c785d421298283ee9212cb2d4663664122fea120c43a0"} Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.044572 4998 scope.go:117] "RemoveContainer" containerID="682763cd65cf4804dc534645f561db0b3daf020b18f1b16c9b083082fae8e47d" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.044889 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.087711 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.096328 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.108047 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 16:27:18 crc kubenswrapper[4998]: E1203 16:27:18.108467 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37109c33-9436-40d5-aa8f-bb7e7c3a0f83" containerName="nova-scheduler-scheduler" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.108483 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="37109c33-9436-40d5-aa8f-bb7e7c3a0f83" containerName="nova-scheduler-scheduler" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.108640 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="37109c33-9436-40d5-aa8f-bb7e7c3a0f83" containerName="nova-scheduler-scheduler" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.109385 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.121079 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.128795 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.292958 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlv55\" (UniqueName: \"kubernetes.io/projected/2b29e862-8b36-45b1-a23a-534083def903-kube-api-access-mlv55\") pod \"nova-scheduler-0\" (UID: \"2b29e862-8b36-45b1-a23a-534083def903\") " pod="openstack/nova-scheduler-0" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.293038 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b29e862-8b36-45b1-a23a-534083def903-config-data\") pod \"nova-scheduler-0\" (UID: \"2b29e862-8b36-45b1-a23a-534083def903\") " pod="openstack/nova-scheduler-0" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.293240 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b29e862-8b36-45b1-a23a-534083def903-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2b29e862-8b36-45b1-a23a-534083def903\") " pod="openstack/nova-scheduler-0" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.395933 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b29e862-8b36-45b1-a23a-534083def903-config-data\") pod \"nova-scheduler-0\" (UID: \"2b29e862-8b36-45b1-a23a-534083def903\") " pod="openstack/nova-scheduler-0" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.396006 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b29e862-8b36-45b1-a23a-534083def903-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2b29e862-8b36-45b1-a23a-534083def903\") " pod="openstack/nova-scheduler-0" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.396137 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlv55\" (UniqueName: \"kubernetes.io/projected/2b29e862-8b36-45b1-a23a-534083def903-kube-api-access-mlv55\") pod \"nova-scheduler-0\" (UID: \"2b29e862-8b36-45b1-a23a-534083def903\") " pod="openstack/nova-scheduler-0" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.399682 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b29e862-8b36-45b1-a23a-534083def903-config-data\") pod \"nova-scheduler-0\" (UID: \"2b29e862-8b36-45b1-a23a-534083def903\") " pod="openstack/nova-scheduler-0" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.400940 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b29e862-8b36-45b1-a23a-534083def903-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2b29e862-8b36-45b1-a23a-534083def903\") " pod="openstack/nova-scheduler-0" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.411437 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlv55\" (UniqueName: \"kubernetes.io/projected/2b29e862-8b36-45b1-a23a-534083def903-kube-api-access-mlv55\") pod \"nova-scheduler-0\" (UID: \"2b29e862-8b36-45b1-a23a-534083def903\") " pod="openstack/nova-scheduler-0" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.424646 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.510501 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vvjng" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.700647 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86d54b38-8913-4d18-92ee-490aceef1677-scripts\") pod \"86d54b38-8913-4d18-92ee-490aceef1677\" (UID: \"86d54b38-8913-4d18-92ee-490aceef1677\") " Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.701039 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d54b38-8913-4d18-92ee-490aceef1677-combined-ca-bundle\") pod \"86d54b38-8913-4d18-92ee-490aceef1677\" (UID: \"86d54b38-8913-4d18-92ee-490aceef1677\") " Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.701193 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgtbd\" (UniqueName: \"kubernetes.io/projected/86d54b38-8913-4d18-92ee-490aceef1677-kube-api-access-vgtbd\") pod \"86d54b38-8913-4d18-92ee-490aceef1677\" (UID: \"86d54b38-8913-4d18-92ee-490aceef1677\") " Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.701273 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86d54b38-8913-4d18-92ee-490aceef1677-config-data\") pod \"86d54b38-8913-4d18-92ee-490aceef1677\" (UID: \"86d54b38-8913-4d18-92ee-490aceef1677\") " Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.718195 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86d54b38-8913-4d18-92ee-490aceef1677-scripts" (OuterVolumeSpecName: "scripts") pod "86d54b38-8913-4d18-92ee-490aceef1677" (UID: "86d54b38-8913-4d18-92ee-490aceef1677"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.718374 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86d54b38-8913-4d18-92ee-490aceef1677-kube-api-access-vgtbd" (OuterVolumeSpecName: "kube-api-access-vgtbd") pod "86d54b38-8913-4d18-92ee-490aceef1677" (UID: "86d54b38-8913-4d18-92ee-490aceef1677"). InnerVolumeSpecName "kube-api-access-vgtbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.749508 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86d54b38-8913-4d18-92ee-490aceef1677-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "86d54b38-8913-4d18-92ee-490aceef1677" (UID: "86d54b38-8913-4d18-92ee-490aceef1677"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.760933 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86d54b38-8913-4d18-92ee-490aceef1677-config-data" (OuterVolumeSpecName: "config-data") pod "86d54b38-8913-4d18-92ee-490aceef1677" (UID: "86d54b38-8913-4d18-92ee-490aceef1677"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.807361 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86d54b38-8913-4d18-92ee-490aceef1677-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.807399 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86d54b38-8913-4d18-92ee-490aceef1677-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.807413 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgtbd\" (UniqueName: \"kubernetes.io/projected/86d54b38-8913-4d18-92ee-490aceef1677-kube-api-access-vgtbd\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.807425 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86d54b38-8913-4d18-92ee-490aceef1677-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:18 crc kubenswrapper[4998]: I1203 16:27:18.924176 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 16:27:18 crc kubenswrapper[4998]: W1203 16:27:18.936739 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b29e862_8b36_45b1_a23a_534083def903.slice/crio-be2e6757ade78923f7147fcf6c330255631097e98a7ecb255e15dce22120b142 WatchSource:0}: Error finding container be2e6757ade78923f7147fcf6c330255631097e98a7ecb255e15dce22120b142: Status 404 returned error can't find the container with id be2e6757ade78923f7147fcf6c330255631097e98a7ecb255e15dce22120b142 Dec 03 16:27:19 crc kubenswrapper[4998]: I1203 16:27:19.060950 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2b29e862-8b36-45b1-a23a-534083def903","Type":"ContainerStarted","Data":"be2e6757ade78923f7147fcf6c330255631097e98a7ecb255e15dce22120b142"} Dec 03 16:27:19 crc kubenswrapper[4998]: I1203 16:27:19.080377 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5","Type":"ContainerStarted","Data":"872657cb4d2b5e6f6cb1eeab3773bb0cad42991d86cc4d4d3c03d3aeae30eb02"} Dec 03 16:27:19 crc kubenswrapper[4998]: I1203 16:27:19.080416 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5","Type":"ContainerStarted","Data":"6f8e7e8f2329d79fc5b615cf264882592e828b8c5cbc63688e35c28d8af30f5a"} Dec 03 16:27:19 crc kubenswrapper[4998]: I1203 16:27:19.098373 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vvjng" event={"ID":"86d54b38-8913-4d18-92ee-490aceef1677","Type":"ContainerDied","Data":"a4321b32108f5fdfe6edbcb98b07cbdfb20c6f7fb29e9c891a769463a10e12b5"} Dec 03 16:27:19 crc kubenswrapper[4998]: I1203 16:27:19.098410 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4321b32108f5fdfe6edbcb98b07cbdfb20c6f7fb29e9c891a769463a10e12b5" Dec 03 16:27:19 crc kubenswrapper[4998]: I1203 16:27:19.098494 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vvjng" Dec 03 16:27:19 crc kubenswrapper[4998]: I1203 16:27:19.135663 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.135641577 podStartE2EDuration="2.135641577s" podCreationTimestamp="2025-12-03 16:27:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:27:19.106726223 +0000 UTC m=+1417.718426446" watchObservedRunningTime="2025-12-03 16:27:19.135641577 +0000 UTC m=+1417.747341800" Dec 03 16:27:19 crc kubenswrapper[4998]: I1203 16:27:19.182811 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 16:27:19 crc kubenswrapper[4998]: E1203 16:27:19.191033 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86d54b38-8913-4d18-92ee-490aceef1677" containerName="nova-cell1-conductor-db-sync" Dec 03 16:27:19 crc kubenswrapper[4998]: I1203 16:27:19.191108 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="86d54b38-8913-4d18-92ee-490aceef1677" containerName="nova-cell1-conductor-db-sync" Dec 03 16:27:19 crc kubenswrapper[4998]: I1203 16:27:19.191422 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="86d54b38-8913-4d18-92ee-490aceef1677" containerName="nova-cell1-conductor-db-sync" Dec 03 16:27:19 crc kubenswrapper[4998]: I1203 16:27:19.192207 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 16:27:19 crc kubenswrapper[4998]: I1203 16:27:19.199529 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 16:27:19 crc kubenswrapper[4998]: I1203 16:27:19.208389 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 16:27:19 crc kubenswrapper[4998]: I1203 16:27:19.316435 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a0de9df-91b8-4cdb-a6a0-46d7f8539b1b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2a0de9df-91b8-4cdb-a6a0-46d7f8539b1b\") " pod="openstack/nova-cell1-conductor-0" Dec 03 16:27:19 crc kubenswrapper[4998]: I1203 16:27:19.316571 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a0de9df-91b8-4cdb-a6a0-46d7f8539b1b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2a0de9df-91b8-4cdb-a6a0-46d7f8539b1b\") " pod="openstack/nova-cell1-conductor-0" Dec 03 16:27:19 crc kubenswrapper[4998]: I1203 16:27:19.316605 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtzvs\" (UniqueName: \"kubernetes.io/projected/2a0de9df-91b8-4cdb-a6a0-46d7f8539b1b-kube-api-access-vtzvs\") pod \"nova-cell1-conductor-0\" (UID: \"2a0de9df-91b8-4cdb-a6a0-46d7f8539b1b\") " pod="openstack/nova-cell1-conductor-0" Dec 03 16:27:19 crc kubenswrapper[4998]: I1203 16:27:19.418671 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a0de9df-91b8-4cdb-a6a0-46d7f8539b1b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2a0de9df-91b8-4cdb-a6a0-46d7f8539b1b\") " pod="openstack/nova-cell1-conductor-0" Dec 03 16:27:19 crc kubenswrapper[4998]: I1203 16:27:19.419024 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtzvs\" (UniqueName: \"kubernetes.io/projected/2a0de9df-91b8-4cdb-a6a0-46d7f8539b1b-kube-api-access-vtzvs\") pod \"nova-cell1-conductor-0\" (UID: \"2a0de9df-91b8-4cdb-a6a0-46d7f8539b1b\") " pod="openstack/nova-cell1-conductor-0" Dec 03 16:27:19 crc kubenswrapper[4998]: I1203 16:27:19.419094 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a0de9df-91b8-4cdb-a6a0-46d7f8539b1b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2a0de9df-91b8-4cdb-a6a0-46d7f8539b1b\") " pod="openstack/nova-cell1-conductor-0" Dec 03 16:27:19 crc kubenswrapper[4998]: I1203 16:27:19.423469 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a0de9df-91b8-4cdb-a6a0-46d7f8539b1b-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2a0de9df-91b8-4cdb-a6a0-46d7f8539b1b\") " pod="openstack/nova-cell1-conductor-0" Dec 03 16:27:19 crc kubenswrapper[4998]: I1203 16:27:19.424956 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a0de9df-91b8-4cdb-a6a0-46d7f8539b1b-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2a0de9df-91b8-4cdb-a6a0-46d7f8539b1b\") " pod="openstack/nova-cell1-conductor-0" Dec 03 16:27:19 crc kubenswrapper[4998]: I1203 16:27:19.449929 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtzvs\" (UniqueName: \"kubernetes.io/projected/2a0de9df-91b8-4cdb-a6a0-46d7f8539b1b-kube-api-access-vtzvs\") pod \"nova-cell1-conductor-0\" (UID: \"2a0de9df-91b8-4cdb-a6a0-46d7f8539b1b\") " pod="openstack/nova-cell1-conductor-0" Dec 03 16:27:19 crc kubenswrapper[4998]: I1203 16:27:19.566795 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 16:27:19 crc kubenswrapper[4998]: I1203 16:27:19.699790 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37109c33-9436-40d5-aa8f-bb7e7c3a0f83" path="/var/lib/kubelet/pods/37109c33-9436-40d5-aa8f-bb7e7c3a0f83/volumes" Dec 03 16:27:20 crc kubenswrapper[4998]: I1203 16:27:20.080478 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 16:27:20 crc kubenswrapper[4998]: I1203 16:27:20.122587 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2b29e862-8b36-45b1-a23a-534083def903","Type":"ContainerStarted","Data":"f7b11db1f404c817ddc5d1b94ead20d1cc0cac2133efa6d55366db7bf306214f"} Dec 03 16:27:20 crc kubenswrapper[4998]: I1203 16:27:20.132031 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2a0de9df-91b8-4cdb-a6a0-46d7f8539b1b","Type":"ContainerStarted","Data":"f33763e8510cb9fe632a2e71a614a02d7d0ac5810b63334d9a42e0ad8eb83fab"} Dec 03 16:27:20 crc kubenswrapper[4998]: I1203 16:27:20.143615 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.143598214 podStartE2EDuration="2.143598214s" podCreationTimestamp="2025-12-03 16:27:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:27:20.138507529 +0000 UTC m=+1418.750207752" watchObservedRunningTime="2025-12-03 16:27:20.143598214 +0000 UTC m=+1418.755298437" Dec 03 16:27:20 crc kubenswrapper[4998]: I1203 16:27:20.289092 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 16:27:20 crc kubenswrapper[4998]: I1203 16:27:20.289133 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 16:27:20 crc kubenswrapper[4998]: I1203 16:27:20.615629 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 16:27:20 crc kubenswrapper[4998]: I1203 16:27:20.616129 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="f9abb1de-7121-4de9-833a-9f1ab889b8ff" containerName="kube-state-metrics" containerID="cri-o://3a0b0d4f0a8d7f25f121dff94ba7f15e27e33f0d83a805c785800edd09699873" gracePeriod=30 Dec 03 16:27:21 crc kubenswrapper[4998]: I1203 16:27:21.142867 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2a0de9df-91b8-4cdb-a6a0-46d7f8539b1b","Type":"ContainerStarted","Data":"55be3134c8247f7b9bf9918c3bbcb71f723aea3fc8781ab1f3e050cf30109ee1"} Dec 03 16:27:21 crc kubenswrapper[4998]: I1203 16:27:21.144179 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 03 16:27:21 crc kubenswrapper[4998]: I1203 16:27:21.147939 4998 generic.go:334] "Generic (PLEG): container finished" podID="f9abb1de-7121-4de9-833a-9f1ab889b8ff" containerID="3a0b0d4f0a8d7f25f121dff94ba7f15e27e33f0d83a805c785800edd09699873" exitCode=2 Dec 03 16:27:21 crc kubenswrapper[4998]: I1203 16:27:21.148296 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f9abb1de-7121-4de9-833a-9f1ab889b8ff","Type":"ContainerDied","Data":"3a0b0d4f0a8d7f25f121dff94ba7f15e27e33f0d83a805c785800edd09699873"} Dec 03 16:27:21 crc kubenswrapper[4998]: I1203 16:27:21.148362 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f9abb1de-7121-4de9-833a-9f1ab889b8ff","Type":"ContainerDied","Data":"dc5aac4b616dd2f5fbb9ca9a9c8ee062f5835f56ad689e0c9da7e0507b7a8a6f"} Dec 03 16:27:21 crc kubenswrapper[4998]: I1203 16:27:21.148378 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc5aac4b616dd2f5fbb9ca9a9c8ee062f5835f56ad689e0c9da7e0507b7a8a6f" Dec 03 16:27:21 crc kubenswrapper[4998]: I1203 16:27:21.162368 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.162352409 podStartE2EDuration="2.162352409s" podCreationTimestamp="2025-12-03 16:27:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:27:21.159980414 +0000 UTC m=+1419.771680637" watchObservedRunningTime="2025-12-03 16:27:21.162352409 +0000 UTC m=+1419.774052632" Dec 03 16:27:21 crc kubenswrapper[4998]: I1203 16:27:21.195340 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 16:27:21 crc kubenswrapper[4998]: I1203 16:27:21.355280 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xz9qq\" (UniqueName: \"kubernetes.io/projected/f9abb1de-7121-4de9-833a-9f1ab889b8ff-kube-api-access-xz9qq\") pod \"f9abb1de-7121-4de9-833a-9f1ab889b8ff\" (UID: \"f9abb1de-7121-4de9-833a-9f1ab889b8ff\") " Dec 03 16:27:21 crc kubenswrapper[4998]: I1203 16:27:21.388952 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9abb1de-7121-4de9-833a-9f1ab889b8ff-kube-api-access-xz9qq" (OuterVolumeSpecName: "kube-api-access-xz9qq") pod "f9abb1de-7121-4de9-833a-9f1ab889b8ff" (UID: "f9abb1de-7121-4de9-833a-9f1ab889b8ff"). InnerVolumeSpecName "kube-api-access-xz9qq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:27:21 crc kubenswrapper[4998]: I1203 16:27:21.458612 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xz9qq\" (UniqueName: \"kubernetes.io/projected/f9abb1de-7121-4de9-833a-9f1ab889b8ff-kube-api-access-xz9qq\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.157557 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.188103 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.202663 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.214654 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 16:27:22 crc kubenswrapper[4998]: E1203 16:27:22.215415 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9abb1de-7121-4de9-833a-9f1ab889b8ff" containerName="kube-state-metrics" Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.215445 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9abb1de-7121-4de9-833a-9f1ab889b8ff" containerName="kube-state-metrics" Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.215855 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9abb1de-7121-4de9-833a-9f1ab889b8ff" containerName="kube-state-metrics" Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.217021 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.220476 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.223671 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.225280 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.378662 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df7dec-fa0b-4b79-8e59-5298650093e5-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"84df7dec-fa0b-4b79-8e59-5298650093e5\") " pod="openstack/kube-state-metrics-0" Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.378723 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9ncs\" (UniqueName: \"kubernetes.io/projected/84df7dec-fa0b-4b79-8e59-5298650093e5-kube-api-access-r9ncs\") pod \"kube-state-metrics-0\" (UID: \"84df7dec-fa0b-4b79-8e59-5298650093e5\") " pod="openstack/kube-state-metrics-0" Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.378861 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/84df7dec-fa0b-4b79-8e59-5298650093e5-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"84df7dec-fa0b-4b79-8e59-5298650093e5\") " pod="openstack/kube-state-metrics-0" Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.379029 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/84df7dec-fa0b-4b79-8e59-5298650093e5-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"84df7dec-fa0b-4b79-8e59-5298650093e5\") " pod="openstack/kube-state-metrics-0" Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.482571 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/84df7dec-fa0b-4b79-8e59-5298650093e5-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"84df7dec-fa0b-4b79-8e59-5298650093e5\") " pod="openstack/kube-state-metrics-0" Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.482706 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df7dec-fa0b-4b79-8e59-5298650093e5-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"84df7dec-fa0b-4b79-8e59-5298650093e5\") " pod="openstack/kube-state-metrics-0" Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.482733 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9ncs\" (UniqueName: \"kubernetes.io/projected/84df7dec-fa0b-4b79-8e59-5298650093e5-kube-api-access-r9ncs\") pod \"kube-state-metrics-0\" (UID: \"84df7dec-fa0b-4b79-8e59-5298650093e5\") " pod="openstack/kube-state-metrics-0" Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.482830 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/84df7dec-fa0b-4b79-8e59-5298650093e5-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"84df7dec-fa0b-4b79-8e59-5298650093e5\") " pod="openstack/kube-state-metrics-0" Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.488953 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84df7dec-fa0b-4b79-8e59-5298650093e5-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"84df7dec-fa0b-4b79-8e59-5298650093e5\") " pod="openstack/kube-state-metrics-0" Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.491505 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/84df7dec-fa0b-4b79-8e59-5298650093e5-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"84df7dec-fa0b-4b79-8e59-5298650093e5\") " pod="openstack/kube-state-metrics-0" Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.501857 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/84df7dec-fa0b-4b79-8e59-5298650093e5-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"84df7dec-fa0b-4b79-8e59-5298650093e5\") " pod="openstack/kube-state-metrics-0" Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.507442 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9ncs\" (UniqueName: \"kubernetes.io/projected/84df7dec-fa0b-4b79-8e59-5298650093e5-kube-api-access-r9ncs\") pod \"kube-state-metrics-0\" (UID: \"84df7dec-fa0b-4b79-8e59-5298650093e5\") " pod="openstack/kube-state-metrics-0" Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.535342 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.848774 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.849266 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac504098-7519-43e2-80ba-b3da6061efc4" containerName="ceilometer-central-agent" containerID="cri-o://b5a5d7ac8282290b1e195561c9e717c4160d4d017b164b09665421b9d7d39472" gracePeriod=30 Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.849341 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac504098-7519-43e2-80ba-b3da6061efc4" containerName="proxy-httpd" containerID="cri-o://a586c5395556d7a52f9e901112e597939bedb0c521f42bc9cbf5d4544587c489" gracePeriod=30 Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.849384 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac504098-7519-43e2-80ba-b3da6061efc4" containerName="sg-core" containerID="cri-o://45258dafb53d8eba634abe2f783c47b774b6165cbb5a248abea8ae51dd949cbb" gracePeriod=30 Dec 03 16:27:22 crc kubenswrapper[4998]: I1203 16:27:22.849380 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac504098-7519-43e2-80ba-b3da6061efc4" containerName="ceilometer-notification-agent" containerID="cri-o://264ea4221f3a7fdc459a2d05f76fa01f2aba5f42bd13e4c928869da5d4cf8838" gracePeriod=30 Dec 03 16:27:23 crc kubenswrapper[4998]: I1203 16:27:23.108362 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 16:27:23 crc kubenswrapper[4998]: W1203 16:27:23.109530 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84df7dec_fa0b_4b79_8e59_5298650093e5.slice/crio-b40d1631eee7891183beada5ee06883a10afa2f90fecdcd06bd038096f940e51 WatchSource:0}: Error finding container b40d1631eee7891183beada5ee06883a10afa2f90fecdcd06bd038096f940e51: Status 404 returned error can't find the container with id b40d1631eee7891183beada5ee06883a10afa2f90fecdcd06bd038096f940e51 Dec 03 16:27:23 crc kubenswrapper[4998]: I1203 16:27:23.169794 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"84df7dec-fa0b-4b79-8e59-5298650093e5","Type":"ContainerStarted","Data":"b40d1631eee7891183beada5ee06883a10afa2f90fecdcd06bd038096f940e51"} Dec 03 16:27:23 crc kubenswrapper[4998]: I1203 16:27:23.172808 4998 generic.go:334] "Generic (PLEG): container finished" podID="ac504098-7519-43e2-80ba-b3da6061efc4" containerID="a586c5395556d7a52f9e901112e597939bedb0c521f42bc9cbf5d4544587c489" exitCode=0 Dec 03 16:27:23 crc kubenswrapper[4998]: I1203 16:27:23.172849 4998 generic.go:334] "Generic (PLEG): container finished" podID="ac504098-7519-43e2-80ba-b3da6061efc4" containerID="45258dafb53d8eba634abe2f783c47b774b6165cbb5a248abea8ae51dd949cbb" exitCode=2 Dec 03 16:27:23 crc kubenswrapper[4998]: I1203 16:27:23.172894 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac504098-7519-43e2-80ba-b3da6061efc4","Type":"ContainerDied","Data":"a586c5395556d7a52f9e901112e597939bedb0c521f42bc9cbf5d4544587c489"} Dec 03 16:27:23 crc kubenswrapper[4998]: I1203 16:27:23.172926 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac504098-7519-43e2-80ba-b3da6061efc4","Type":"ContainerDied","Data":"45258dafb53d8eba634abe2f783c47b774b6165cbb5a248abea8ae51dd949cbb"} Dec 03 16:27:23 crc kubenswrapper[4998]: I1203 16:27:23.425704 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 16:27:23 crc kubenswrapper[4998]: I1203 16:27:23.688957 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9abb1de-7121-4de9-833a-9f1ab889b8ff" path="/var/lib/kubelet/pods/f9abb1de-7121-4de9-833a-9f1ab889b8ff/volumes" Dec 03 16:27:24 crc kubenswrapper[4998]: I1203 16:27:24.186895 4998 generic.go:334] "Generic (PLEG): container finished" podID="ac504098-7519-43e2-80ba-b3da6061efc4" containerID="b5a5d7ac8282290b1e195561c9e717c4160d4d017b164b09665421b9d7d39472" exitCode=0 Dec 03 16:27:24 crc kubenswrapper[4998]: I1203 16:27:24.186949 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac504098-7519-43e2-80ba-b3da6061efc4","Type":"ContainerDied","Data":"b5a5d7ac8282290b1e195561c9e717c4160d4d017b164b09665421b9d7d39472"} Dec 03 16:27:25 crc kubenswrapper[4998]: I1203 16:27:25.289700 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 16:27:25 crc kubenswrapper[4998]: I1203 16:27:25.290225 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 16:27:26 crc kubenswrapper[4998]: I1203 16:27:26.305887 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="cd010875-398a-48fc-8fa5-569126c4f67f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.214:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 16:27:26 crc kubenswrapper[4998]: I1203 16:27:26.305940 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="cd010875-398a-48fc-8fa5-569126c4f67f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.214:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 16:27:27 crc kubenswrapper[4998]: I1203 16:27:27.117013 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:27:27 crc kubenswrapper[4998]: I1203 16:27:27.117081 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:27:27 crc kubenswrapper[4998]: I1203 16:27:27.457348 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 16:27:27 crc kubenswrapper[4998]: I1203 16:27:27.457664 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 16:27:28 crc kubenswrapper[4998]: I1203 16:27:28.232106 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"84df7dec-fa0b-4b79-8e59-5298650093e5","Type":"ContainerStarted","Data":"f9c6e2fbfa200fb52a7d035e6f7f8eed6cbd874b15324288fa075e90af63ff05"} Dec 03 16:27:28 crc kubenswrapper[4998]: I1203 16:27:28.232935 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 16:27:28 crc kubenswrapper[4998]: I1203 16:27:28.255244 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.301410719 podStartE2EDuration="6.255219091s" podCreationTimestamp="2025-12-03 16:27:22 +0000 UTC" firstStartedPulling="2025-12-03 16:27:23.112416586 +0000 UTC m=+1421.724116809" lastFinishedPulling="2025-12-03 16:27:27.066224958 +0000 UTC m=+1425.677925181" observedRunningTime="2025-12-03 16:27:28.249938363 +0000 UTC m=+1426.861638596" watchObservedRunningTime="2025-12-03 16:27:28.255219091 +0000 UTC m=+1426.866919314" Dec 03 16:27:28 crc kubenswrapper[4998]: I1203 16:27:28.425152 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 16:27:28 crc kubenswrapper[4998]: I1203 16:27:28.459139 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 16:27:28 crc kubenswrapper[4998]: I1203 16:27:28.541406 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.215:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 16:27:28 crc kubenswrapper[4998]: I1203 16:27:28.541406 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.215:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.045418 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.169715 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac504098-7519-43e2-80ba-b3da6061efc4-combined-ca-bundle\") pod \"ac504098-7519-43e2-80ba-b3da6061efc4\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.169861 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac504098-7519-43e2-80ba-b3da6061efc4-run-httpd\") pod \"ac504098-7519-43e2-80ba-b3da6061efc4\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.169936 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac504098-7519-43e2-80ba-b3da6061efc4-scripts\") pod \"ac504098-7519-43e2-80ba-b3da6061efc4\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.169983 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac504098-7519-43e2-80ba-b3da6061efc4-config-data\") pod \"ac504098-7519-43e2-80ba-b3da6061efc4\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.170075 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac504098-7519-43e2-80ba-b3da6061efc4-sg-core-conf-yaml\") pod \"ac504098-7519-43e2-80ba-b3da6061efc4\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.170110 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df4bp\" (UniqueName: \"kubernetes.io/projected/ac504098-7519-43e2-80ba-b3da6061efc4-kube-api-access-df4bp\") pod \"ac504098-7519-43e2-80ba-b3da6061efc4\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.170228 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac504098-7519-43e2-80ba-b3da6061efc4-log-httpd\") pod \"ac504098-7519-43e2-80ba-b3da6061efc4\" (UID: \"ac504098-7519-43e2-80ba-b3da6061efc4\") " Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.171638 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac504098-7519-43e2-80ba-b3da6061efc4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ac504098-7519-43e2-80ba-b3da6061efc4" (UID: "ac504098-7519-43e2-80ba-b3da6061efc4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.172105 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac504098-7519-43e2-80ba-b3da6061efc4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ac504098-7519-43e2-80ba-b3da6061efc4" (UID: "ac504098-7519-43e2-80ba-b3da6061efc4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.178306 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac504098-7519-43e2-80ba-b3da6061efc4-scripts" (OuterVolumeSpecName: "scripts") pod "ac504098-7519-43e2-80ba-b3da6061efc4" (UID: "ac504098-7519-43e2-80ba-b3da6061efc4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.180633 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac504098-7519-43e2-80ba-b3da6061efc4-kube-api-access-df4bp" (OuterVolumeSpecName: "kube-api-access-df4bp") pod "ac504098-7519-43e2-80ba-b3da6061efc4" (UID: "ac504098-7519-43e2-80ba-b3da6061efc4"). InnerVolumeSpecName "kube-api-access-df4bp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.207902 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac504098-7519-43e2-80ba-b3da6061efc4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ac504098-7519-43e2-80ba-b3da6061efc4" (UID: "ac504098-7519-43e2-80ba-b3da6061efc4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.246228 4998 generic.go:334] "Generic (PLEG): container finished" podID="ac504098-7519-43e2-80ba-b3da6061efc4" containerID="264ea4221f3a7fdc459a2d05f76fa01f2aba5f42bd13e4c928869da5d4cf8838" exitCode=0 Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.247170 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.247189 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac504098-7519-43e2-80ba-b3da6061efc4","Type":"ContainerDied","Data":"264ea4221f3a7fdc459a2d05f76fa01f2aba5f42bd13e4c928869da5d4cf8838"} Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.247311 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac504098-7519-43e2-80ba-b3da6061efc4","Type":"ContainerDied","Data":"3ec0705c5c1d063e67a3907edad29bc88b6cb739a6e44348cbe69518694a2618"} Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.247354 4998 scope.go:117] "RemoveContainer" containerID="a586c5395556d7a52f9e901112e597939bedb0c521f42bc9cbf5d4544587c489" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.259328 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac504098-7519-43e2-80ba-b3da6061efc4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac504098-7519-43e2-80ba-b3da6061efc4" (UID: "ac504098-7519-43e2-80ba-b3da6061efc4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.272917 4998 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac504098-7519-43e2-80ba-b3da6061efc4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.272949 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df4bp\" (UniqueName: \"kubernetes.io/projected/ac504098-7519-43e2-80ba-b3da6061efc4-kube-api-access-df4bp\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.272963 4998 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac504098-7519-43e2-80ba-b3da6061efc4-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.272972 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac504098-7519-43e2-80ba-b3da6061efc4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.272982 4998 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac504098-7519-43e2-80ba-b3da6061efc4-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.272995 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac504098-7519-43e2-80ba-b3da6061efc4-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.285571 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.305757 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac504098-7519-43e2-80ba-b3da6061efc4-config-data" (OuterVolumeSpecName: "config-data") pod "ac504098-7519-43e2-80ba-b3da6061efc4" (UID: "ac504098-7519-43e2-80ba-b3da6061efc4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.306852 4998 scope.go:117] "RemoveContainer" containerID="45258dafb53d8eba634abe2f783c47b774b6165cbb5a248abea8ae51dd949cbb" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.347509 4998 scope.go:117] "RemoveContainer" containerID="264ea4221f3a7fdc459a2d05f76fa01f2aba5f42bd13e4c928869da5d4cf8838" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.375056 4998 scope.go:117] "RemoveContainer" containerID="b5a5d7ac8282290b1e195561c9e717c4160d4d017b164b09665421b9d7d39472" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.377402 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac504098-7519-43e2-80ba-b3da6061efc4-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.400555 4998 scope.go:117] "RemoveContainer" containerID="a586c5395556d7a52f9e901112e597939bedb0c521f42bc9cbf5d4544587c489" Dec 03 16:27:29 crc kubenswrapper[4998]: E1203 16:27:29.404679 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a586c5395556d7a52f9e901112e597939bedb0c521f42bc9cbf5d4544587c489\": container with ID starting with a586c5395556d7a52f9e901112e597939bedb0c521f42bc9cbf5d4544587c489 not found: ID does not exist" containerID="a586c5395556d7a52f9e901112e597939bedb0c521f42bc9cbf5d4544587c489" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.404740 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a586c5395556d7a52f9e901112e597939bedb0c521f42bc9cbf5d4544587c489"} err="failed to get container status \"a586c5395556d7a52f9e901112e597939bedb0c521f42bc9cbf5d4544587c489\": rpc error: code = NotFound desc = could not find container \"a586c5395556d7a52f9e901112e597939bedb0c521f42bc9cbf5d4544587c489\": container with ID starting with a586c5395556d7a52f9e901112e597939bedb0c521f42bc9cbf5d4544587c489 not found: ID does not exist" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.404787 4998 scope.go:117] "RemoveContainer" containerID="45258dafb53d8eba634abe2f783c47b774b6165cbb5a248abea8ae51dd949cbb" Dec 03 16:27:29 crc kubenswrapper[4998]: E1203 16:27:29.405697 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45258dafb53d8eba634abe2f783c47b774b6165cbb5a248abea8ae51dd949cbb\": container with ID starting with 45258dafb53d8eba634abe2f783c47b774b6165cbb5a248abea8ae51dd949cbb not found: ID does not exist" containerID="45258dafb53d8eba634abe2f783c47b774b6165cbb5a248abea8ae51dd949cbb" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.405737 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45258dafb53d8eba634abe2f783c47b774b6165cbb5a248abea8ae51dd949cbb"} err="failed to get container status \"45258dafb53d8eba634abe2f783c47b774b6165cbb5a248abea8ae51dd949cbb\": rpc error: code = NotFound desc = could not find container \"45258dafb53d8eba634abe2f783c47b774b6165cbb5a248abea8ae51dd949cbb\": container with ID starting with 45258dafb53d8eba634abe2f783c47b774b6165cbb5a248abea8ae51dd949cbb not found: ID does not exist" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.405771 4998 scope.go:117] "RemoveContainer" containerID="264ea4221f3a7fdc459a2d05f76fa01f2aba5f42bd13e4c928869da5d4cf8838" Dec 03 16:27:29 crc kubenswrapper[4998]: E1203 16:27:29.406108 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"264ea4221f3a7fdc459a2d05f76fa01f2aba5f42bd13e4c928869da5d4cf8838\": container with ID starting with 264ea4221f3a7fdc459a2d05f76fa01f2aba5f42bd13e4c928869da5d4cf8838 not found: ID does not exist" containerID="264ea4221f3a7fdc459a2d05f76fa01f2aba5f42bd13e4c928869da5d4cf8838" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.406148 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"264ea4221f3a7fdc459a2d05f76fa01f2aba5f42bd13e4c928869da5d4cf8838"} err="failed to get container status \"264ea4221f3a7fdc459a2d05f76fa01f2aba5f42bd13e4c928869da5d4cf8838\": rpc error: code = NotFound desc = could not find container \"264ea4221f3a7fdc459a2d05f76fa01f2aba5f42bd13e4c928869da5d4cf8838\": container with ID starting with 264ea4221f3a7fdc459a2d05f76fa01f2aba5f42bd13e4c928869da5d4cf8838 not found: ID does not exist" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.406170 4998 scope.go:117] "RemoveContainer" containerID="b5a5d7ac8282290b1e195561c9e717c4160d4d017b164b09665421b9d7d39472" Dec 03 16:27:29 crc kubenswrapper[4998]: E1203 16:27:29.406776 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5a5d7ac8282290b1e195561c9e717c4160d4d017b164b09665421b9d7d39472\": container with ID starting with b5a5d7ac8282290b1e195561c9e717c4160d4d017b164b09665421b9d7d39472 not found: ID does not exist" containerID="b5a5d7ac8282290b1e195561c9e717c4160d4d017b164b09665421b9d7d39472" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.406803 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5a5d7ac8282290b1e195561c9e717c4160d4d017b164b09665421b9d7d39472"} err="failed to get container status \"b5a5d7ac8282290b1e195561c9e717c4160d4d017b164b09665421b9d7d39472\": rpc error: code = NotFound desc = could not find container \"b5a5d7ac8282290b1e195561c9e717c4160d4d017b164b09665421b9d7d39472\": container with ID starting with b5a5d7ac8282290b1e195561c9e717c4160d4d017b164b09665421b9d7d39472 not found: ID does not exist" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.588164 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.608359 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.615131 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.618138 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:27:29 crc kubenswrapper[4998]: E1203 16:27:29.618733 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac504098-7519-43e2-80ba-b3da6061efc4" containerName="ceilometer-central-agent" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.618767 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac504098-7519-43e2-80ba-b3da6061efc4" containerName="ceilometer-central-agent" Dec 03 16:27:29 crc kubenswrapper[4998]: E1203 16:27:29.618791 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac504098-7519-43e2-80ba-b3da6061efc4" containerName="proxy-httpd" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.618798 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac504098-7519-43e2-80ba-b3da6061efc4" containerName="proxy-httpd" Dec 03 16:27:29 crc kubenswrapper[4998]: E1203 16:27:29.618808 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac504098-7519-43e2-80ba-b3da6061efc4" containerName="ceilometer-notification-agent" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.618815 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac504098-7519-43e2-80ba-b3da6061efc4" containerName="ceilometer-notification-agent" Dec 03 16:27:29 crc kubenswrapper[4998]: E1203 16:27:29.618834 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac504098-7519-43e2-80ba-b3da6061efc4" containerName="sg-core" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.618840 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac504098-7519-43e2-80ba-b3da6061efc4" containerName="sg-core" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.619028 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac504098-7519-43e2-80ba-b3da6061efc4" containerName="sg-core" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.619042 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac504098-7519-43e2-80ba-b3da6061efc4" containerName="ceilometer-notification-agent" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.619065 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac504098-7519-43e2-80ba-b3da6061efc4" containerName="ceilometer-central-agent" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.619078 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac504098-7519-43e2-80ba-b3da6061efc4" containerName="proxy-httpd" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.621516 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.627477 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.627579 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.627723 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.640805 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.700502 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac504098-7519-43e2-80ba-b3da6061efc4" path="/var/lib/kubelet/pods/ac504098-7519-43e2-80ba-b3da6061efc4/volumes" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.792448 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.792587 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-scripts\") pod \"ceilometer-0\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.792638 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q42cn\" (UniqueName: \"kubernetes.io/projected/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-kube-api-access-q42cn\") pod \"ceilometer-0\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.792668 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-run-httpd\") pod \"ceilometer-0\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.792697 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.792717 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-config-data\") pod \"ceilometer-0\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.792838 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-log-httpd\") pod \"ceilometer-0\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.792866 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.894833 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.894951 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-scripts\") pod \"ceilometer-0\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.894988 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q42cn\" (UniqueName: \"kubernetes.io/projected/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-kube-api-access-q42cn\") pod \"ceilometer-0\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.895040 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-run-httpd\") pod \"ceilometer-0\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.895078 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.895101 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-config-data\") pod \"ceilometer-0\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.895177 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-log-httpd\") pod \"ceilometer-0\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.895214 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.895816 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-log-httpd\") pod \"ceilometer-0\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.896123 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-run-httpd\") pod \"ceilometer-0\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.899479 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-config-data\") pod \"ceilometer-0\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.899531 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-scripts\") pod \"ceilometer-0\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.899881 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.913152 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.914323 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q42cn\" (UniqueName: \"kubernetes.io/projected/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-kube-api-access-q42cn\") pod \"ceilometer-0\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.914361 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " pod="openstack/ceilometer-0" Dec 03 16:27:29 crc kubenswrapper[4998]: I1203 16:27:29.946607 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:27:31 crc kubenswrapper[4998]: I1203 16:27:30.343516 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:27:31 crc kubenswrapper[4998]: W1203 16:27:30.346082 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8032c83a_f3f0_4cce_a1f3_7ec9e26ba7af.slice/crio-7369788119e596e8216be3c0c58d630cf883d6dadbb9b5bbd9084112ca939a85 WatchSource:0}: Error finding container 7369788119e596e8216be3c0c58d630cf883d6dadbb9b5bbd9084112ca939a85: Status 404 returned error can't find the container with id 7369788119e596e8216be3c0c58d630cf883d6dadbb9b5bbd9084112ca939a85 Dec 03 16:27:31 crc kubenswrapper[4998]: I1203 16:27:31.288496 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af","Type":"ContainerStarted","Data":"5866a4f0439acbebc83da557dba8915695913a9bbc0b91e588977170748394dd"} Dec 03 16:27:31 crc kubenswrapper[4998]: I1203 16:27:31.289340 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af","Type":"ContainerStarted","Data":"4215d5e99baf226e7f679ab78b4f3a7faf72ab49a17d64c14ea966671e5c53c1"} Dec 03 16:27:31 crc kubenswrapper[4998]: I1203 16:27:31.289355 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af","Type":"ContainerStarted","Data":"7369788119e596e8216be3c0c58d630cf883d6dadbb9b5bbd9084112ca939a85"} Dec 03 16:27:32 crc kubenswrapper[4998]: I1203 16:27:32.308497 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af","Type":"ContainerStarted","Data":"44b9c98072d88268c274fdada0200c7d6641a4b6f7548997a7c9b7bc3ef1a976"} Dec 03 16:27:32 crc kubenswrapper[4998]: I1203 16:27:32.547132 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 16:27:34 crc kubenswrapper[4998]: I1203 16:27:34.339024 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af","Type":"ContainerStarted","Data":"8b7f798da97884f65fa3da02156f3092bac3a18f0058055f6c7a8c2d74333875"} Dec 03 16:27:34 crc kubenswrapper[4998]: I1203 16:27:34.342491 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 16:27:34 crc kubenswrapper[4998]: I1203 16:27:34.392899 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.153542408 podStartE2EDuration="5.392876849s" podCreationTimestamp="2025-12-03 16:27:29 +0000 UTC" firstStartedPulling="2025-12-03 16:27:30.348806983 +0000 UTC m=+1428.960507206" lastFinishedPulling="2025-12-03 16:27:33.588141424 +0000 UTC m=+1432.199841647" observedRunningTime="2025-12-03 16:27:34.376919895 +0000 UTC m=+1432.988620148" watchObservedRunningTime="2025-12-03 16:27:34.392876849 +0000 UTC m=+1433.004577082" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.128338 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.208590 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cfaa0aa-a7ff-4349-ae1f-2403f81041b0-config-data\") pod \"6cfaa0aa-a7ff-4349-ae1f-2403f81041b0\" (UID: \"6cfaa0aa-a7ff-4349-ae1f-2403f81041b0\") " Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.208649 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2m7rw\" (UniqueName: \"kubernetes.io/projected/6cfaa0aa-a7ff-4349-ae1f-2403f81041b0-kube-api-access-2m7rw\") pod \"6cfaa0aa-a7ff-4349-ae1f-2403f81041b0\" (UID: \"6cfaa0aa-a7ff-4349-ae1f-2403f81041b0\") " Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.208935 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cfaa0aa-a7ff-4349-ae1f-2403f81041b0-combined-ca-bundle\") pod \"6cfaa0aa-a7ff-4349-ae1f-2403f81041b0\" (UID: \"6cfaa0aa-a7ff-4349-ae1f-2403f81041b0\") " Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.214991 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cfaa0aa-a7ff-4349-ae1f-2403f81041b0-kube-api-access-2m7rw" (OuterVolumeSpecName: "kube-api-access-2m7rw") pod "6cfaa0aa-a7ff-4349-ae1f-2403f81041b0" (UID: "6cfaa0aa-a7ff-4349-ae1f-2403f81041b0"). InnerVolumeSpecName "kube-api-access-2m7rw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.247780 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cfaa0aa-a7ff-4349-ae1f-2403f81041b0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6cfaa0aa-a7ff-4349-ae1f-2403f81041b0" (UID: "6cfaa0aa-a7ff-4349-ae1f-2403f81041b0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.248375 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cfaa0aa-a7ff-4349-ae1f-2403f81041b0-config-data" (OuterVolumeSpecName: "config-data") pod "6cfaa0aa-a7ff-4349-ae1f-2403f81041b0" (UID: "6cfaa0aa-a7ff-4349-ae1f-2403f81041b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.298327 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.299519 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.305887 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.311318 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cfaa0aa-a7ff-4349-ae1f-2403f81041b0-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.311353 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2m7rw\" (UniqueName: \"kubernetes.io/projected/6cfaa0aa-a7ff-4349-ae1f-2403f81041b0-kube-api-access-2m7rw\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.311362 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cfaa0aa-a7ff-4349-ae1f-2403f81041b0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.352399 4998 generic.go:334] "Generic (PLEG): container finished" podID="6cfaa0aa-a7ff-4349-ae1f-2403f81041b0" containerID="f2306dc70408dbd35479757f2334c590768a431598426f2a974900cde2b177b7" exitCode=137 Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.352456 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.352495 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6cfaa0aa-a7ff-4349-ae1f-2403f81041b0","Type":"ContainerDied","Data":"f2306dc70408dbd35479757f2334c590768a431598426f2a974900cde2b177b7"} Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.352542 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6cfaa0aa-a7ff-4349-ae1f-2403f81041b0","Type":"ContainerDied","Data":"21fbcbb11b108ab59c4a62b60991d51a461bc3ddd9e9731c16f9c82b3f670a06"} Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.352563 4998 scope.go:117] "RemoveContainer" containerID="f2306dc70408dbd35479757f2334c590768a431598426f2a974900cde2b177b7" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.363172 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.401644 4998 scope.go:117] "RemoveContainer" containerID="f2306dc70408dbd35479757f2334c590768a431598426f2a974900cde2b177b7" Dec 03 16:27:35 crc kubenswrapper[4998]: E1203 16:27:35.402185 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2306dc70408dbd35479757f2334c590768a431598426f2a974900cde2b177b7\": container with ID starting with f2306dc70408dbd35479757f2334c590768a431598426f2a974900cde2b177b7 not found: ID does not exist" containerID="f2306dc70408dbd35479757f2334c590768a431598426f2a974900cde2b177b7" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.402239 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2306dc70408dbd35479757f2334c590768a431598426f2a974900cde2b177b7"} err="failed to get container status \"f2306dc70408dbd35479757f2334c590768a431598426f2a974900cde2b177b7\": rpc error: code = NotFound desc = could not find container \"f2306dc70408dbd35479757f2334c590768a431598426f2a974900cde2b177b7\": container with ID starting with f2306dc70408dbd35479757f2334c590768a431598426f2a974900cde2b177b7 not found: ID does not exist" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.419178 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.432466 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.464395 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 16:27:35 crc kubenswrapper[4998]: E1203 16:27:35.465047 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cfaa0aa-a7ff-4349-ae1f-2403f81041b0" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.465073 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cfaa0aa-a7ff-4349-ae1f-2403f81041b0" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.465381 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cfaa0aa-a7ff-4349-ae1f-2403f81041b0" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.466280 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.475891 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.476623 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.479465 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.501410 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.639491 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e064512c-fcb1-4c90-8675-bc9f3991c937-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e064512c-fcb1-4c90-8675-bc9f3991c937\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.639544 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dcv8\" (UniqueName: \"kubernetes.io/projected/e064512c-fcb1-4c90-8675-bc9f3991c937-kube-api-access-5dcv8\") pod \"nova-cell1-novncproxy-0\" (UID: \"e064512c-fcb1-4c90-8675-bc9f3991c937\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.639580 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e064512c-fcb1-4c90-8675-bc9f3991c937-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e064512c-fcb1-4c90-8675-bc9f3991c937\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.639603 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e064512c-fcb1-4c90-8675-bc9f3991c937-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e064512c-fcb1-4c90-8675-bc9f3991c937\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.639885 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e064512c-fcb1-4c90-8675-bc9f3991c937-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e064512c-fcb1-4c90-8675-bc9f3991c937\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.692487 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cfaa0aa-a7ff-4349-ae1f-2403f81041b0" path="/var/lib/kubelet/pods/6cfaa0aa-a7ff-4349-ae1f-2403f81041b0/volumes" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.741997 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dcv8\" (UniqueName: \"kubernetes.io/projected/e064512c-fcb1-4c90-8675-bc9f3991c937-kube-api-access-5dcv8\") pod \"nova-cell1-novncproxy-0\" (UID: \"e064512c-fcb1-4c90-8675-bc9f3991c937\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.742577 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e064512c-fcb1-4c90-8675-bc9f3991c937-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e064512c-fcb1-4c90-8675-bc9f3991c937\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.742606 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e064512c-fcb1-4c90-8675-bc9f3991c937-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e064512c-fcb1-4c90-8675-bc9f3991c937\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.742700 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e064512c-fcb1-4c90-8675-bc9f3991c937-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e064512c-fcb1-4c90-8675-bc9f3991c937\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.742815 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e064512c-fcb1-4c90-8675-bc9f3991c937-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e064512c-fcb1-4c90-8675-bc9f3991c937\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.747158 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e064512c-fcb1-4c90-8675-bc9f3991c937-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e064512c-fcb1-4c90-8675-bc9f3991c937\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.749259 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e064512c-fcb1-4c90-8675-bc9f3991c937-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e064512c-fcb1-4c90-8675-bc9f3991c937\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.749430 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e064512c-fcb1-4c90-8675-bc9f3991c937-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e064512c-fcb1-4c90-8675-bc9f3991c937\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.751573 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e064512c-fcb1-4c90-8675-bc9f3991c937-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e064512c-fcb1-4c90-8675-bc9f3991c937\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.778934 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dcv8\" (UniqueName: \"kubernetes.io/projected/e064512c-fcb1-4c90-8675-bc9f3991c937-kube-api-access-5dcv8\") pod \"nova-cell1-novncproxy-0\" (UID: \"e064512c-fcb1-4c90-8675-bc9f3991c937\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:35 crc kubenswrapper[4998]: I1203 16:27:35.793778 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:36 crc kubenswrapper[4998]: I1203 16:27:36.351239 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 16:27:37 crc kubenswrapper[4998]: I1203 16:27:37.382302 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e064512c-fcb1-4c90-8675-bc9f3991c937","Type":"ContainerStarted","Data":"087e2a06bac70bfb5e26ebd0c134bedf68717ced9913c29780f293bad9c521a6"} Dec 03 16:27:37 crc kubenswrapper[4998]: I1203 16:27:37.382644 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e064512c-fcb1-4c90-8675-bc9f3991c937","Type":"ContainerStarted","Data":"11420861764e555e939cad4bae23b40d8b3ef55afa1953f41d6170c020f00438"} Dec 03 16:27:37 crc kubenswrapper[4998]: I1203 16:27:37.415860 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.415834555 podStartE2EDuration="2.415834555s" podCreationTimestamp="2025-12-03 16:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:27:37.403950561 +0000 UTC m=+1436.015650804" watchObservedRunningTime="2025-12-03 16:27:37.415834555 +0000 UTC m=+1436.027534779" Dec 03 16:27:37 crc kubenswrapper[4998]: I1203 16:27:37.464920 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 16:27:37 crc kubenswrapper[4998]: I1203 16:27:37.465445 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 16:27:37 crc kubenswrapper[4998]: I1203 16:27:37.465508 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 16:27:37 crc kubenswrapper[4998]: I1203 16:27:37.472193 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 16:27:38 crc kubenswrapper[4998]: I1203 16:27:38.393590 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 16:27:38 crc kubenswrapper[4998]: I1203 16:27:38.403655 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 16:27:38 crc kubenswrapper[4998]: I1203 16:27:38.762632 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-fcfbd48f-htqmd"] Dec 03 16:27:38 crc kubenswrapper[4998]: I1203 16:27:38.775523 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" Dec 03 16:27:38 crc kubenswrapper[4998]: I1203 16:27:38.790963 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fcfbd48f-htqmd"] Dec 03 16:27:38 crc kubenswrapper[4998]: I1203 16:27:38.855060 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfbd48f-htqmd\" (UID: \"06704c02-6fcd-4257-bc37-981e2f437db2\") " pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" Dec 03 16:27:38 crc kubenswrapper[4998]: I1203 16:27:38.855383 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfbd48f-htqmd\" (UID: \"06704c02-6fcd-4257-bc37-981e2f437db2\") " pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" Dec 03 16:27:38 crc kubenswrapper[4998]: I1203 16:27:38.855408 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-dns-svc\") pod \"dnsmasq-dns-fcfbd48f-htqmd\" (UID: \"06704c02-6fcd-4257-bc37-981e2f437db2\") " pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" Dec 03 16:27:38 crc kubenswrapper[4998]: I1203 16:27:38.855559 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-config\") pod \"dnsmasq-dns-fcfbd48f-htqmd\" (UID: \"06704c02-6fcd-4257-bc37-981e2f437db2\") " pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" Dec 03 16:27:38 crc kubenswrapper[4998]: I1203 16:27:38.855598 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfbd48f-htqmd\" (UID: \"06704c02-6fcd-4257-bc37-981e2f437db2\") " pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" Dec 03 16:27:38 crc kubenswrapper[4998]: I1203 16:27:38.855661 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkldt\" (UniqueName: \"kubernetes.io/projected/06704c02-6fcd-4257-bc37-981e2f437db2-kube-api-access-bkldt\") pod \"dnsmasq-dns-fcfbd48f-htqmd\" (UID: \"06704c02-6fcd-4257-bc37-981e2f437db2\") " pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" Dec 03 16:27:38 crc kubenswrapper[4998]: I1203 16:27:38.957664 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-dns-svc\") pod \"dnsmasq-dns-fcfbd48f-htqmd\" (UID: \"06704c02-6fcd-4257-bc37-981e2f437db2\") " pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" Dec 03 16:27:38 crc kubenswrapper[4998]: I1203 16:27:38.957699 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfbd48f-htqmd\" (UID: \"06704c02-6fcd-4257-bc37-981e2f437db2\") " pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" Dec 03 16:27:38 crc kubenswrapper[4998]: I1203 16:27:38.957843 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-config\") pod \"dnsmasq-dns-fcfbd48f-htqmd\" (UID: \"06704c02-6fcd-4257-bc37-981e2f437db2\") " pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" Dec 03 16:27:38 crc kubenswrapper[4998]: I1203 16:27:38.957872 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfbd48f-htqmd\" (UID: \"06704c02-6fcd-4257-bc37-981e2f437db2\") " pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" Dec 03 16:27:38 crc kubenswrapper[4998]: I1203 16:27:38.957901 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkldt\" (UniqueName: \"kubernetes.io/projected/06704c02-6fcd-4257-bc37-981e2f437db2-kube-api-access-bkldt\") pod \"dnsmasq-dns-fcfbd48f-htqmd\" (UID: \"06704c02-6fcd-4257-bc37-981e2f437db2\") " pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" Dec 03 16:27:38 crc kubenswrapper[4998]: I1203 16:27:38.957917 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfbd48f-htqmd\" (UID: \"06704c02-6fcd-4257-bc37-981e2f437db2\") " pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" Dec 03 16:27:38 crc kubenswrapper[4998]: I1203 16:27:38.958636 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfbd48f-htqmd\" (UID: \"06704c02-6fcd-4257-bc37-981e2f437db2\") " pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" Dec 03 16:27:38 crc kubenswrapper[4998]: I1203 16:27:38.958632 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-dns-svc\") pod \"dnsmasq-dns-fcfbd48f-htqmd\" (UID: \"06704c02-6fcd-4257-bc37-981e2f437db2\") " pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" Dec 03 16:27:38 crc kubenswrapper[4998]: I1203 16:27:38.959258 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfbd48f-htqmd\" (UID: \"06704c02-6fcd-4257-bc37-981e2f437db2\") " pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" Dec 03 16:27:38 crc kubenswrapper[4998]: I1203 16:27:38.959309 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-config\") pod \"dnsmasq-dns-fcfbd48f-htqmd\" (UID: \"06704c02-6fcd-4257-bc37-981e2f437db2\") " pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" Dec 03 16:27:38 crc kubenswrapper[4998]: I1203 16:27:38.960217 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfbd48f-htqmd\" (UID: \"06704c02-6fcd-4257-bc37-981e2f437db2\") " pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" Dec 03 16:27:39 crc kubenswrapper[4998]: I1203 16:27:39.001934 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkldt\" (UniqueName: \"kubernetes.io/projected/06704c02-6fcd-4257-bc37-981e2f437db2-kube-api-access-bkldt\") pod \"dnsmasq-dns-fcfbd48f-htqmd\" (UID: \"06704c02-6fcd-4257-bc37-981e2f437db2\") " pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" Dec 03 16:27:39 crc kubenswrapper[4998]: I1203 16:27:39.115067 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" Dec 03 16:27:39 crc kubenswrapper[4998]: I1203 16:27:39.703641 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fcfbd48f-htqmd"] Dec 03 16:27:40 crc kubenswrapper[4998]: I1203 16:27:40.413716 4998 generic.go:334] "Generic (PLEG): container finished" podID="06704c02-6fcd-4257-bc37-981e2f437db2" containerID="7f28edb8b8f2a2d11935910fb6e6c424a273b30cbdec046fe5b1fab87d86d9f6" exitCode=0 Dec 03 16:27:40 crc kubenswrapper[4998]: I1203 16:27:40.413858 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" event={"ID":"06704c02-6fcd-4257-bc37-981e2f437db2","Type":"ContainerDied","Data":"7f28edb8b8f2a2d11935910fb6e6c424a273b30cbdec046fe5b1fab87d86d9f6"} Dec 03 16:27:40 crc kubenswrapper[4998]: I1203 16:27:40.413945 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" event={"ID":"06704c02-6fcd-4257-bc37-981e2f437db2","Type":"ContainerStarted","Data":"e5eac56ee0ded262548fa9c7566ed2af682f51ef5f0952365a88ec949dd5a993"} Dec 03 16:27:40 crc kubenswrapper[4998]: I1203 16:27:40.795062 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:41 crc kubenswrapper[4998]: I1203 16:27:41.430941 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" event={"ID":"06704c02-6fcd-4257-bc37-981e2f437db2","Type":"ContainerStarted","Data":"53707713a41b16da049a2c881a7733f518863fb23ce22f106511060a2ba76072"} Dec 03 16:27:41 crc kubenswrapper[4998]: I1203 16:27:41.431154 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" Dec 03 16:27:41 crc kubenswrapper[4998]: I1203 16:27:41.456028 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" podStartSLOduration=3.456006055 podStartE2EDuration="3.456006055s" podCreationTimestamp="2025-12-03 16:27:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:27:41.449397497 +0000 UTC m=+1440.061097720" watchObservedRunningTime="2025-12-03 16:27:41.456006055 +0000 UTC m=+1440.067706288" Dec 03 16:27:41 crc kubenswrapper[4998]: I1203 16:27:41.622714 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 16:27:41 crc kubenswrapper[4998]: I1203 16:27:41.622955 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5" containerName="nova-api-log" containerID="cri-o://6f8e7e8f2329d79fc5b615cf264882592e828b8c5cbc63688e35c28d8af30f5a" gracePeriod=30 Dec 03 16:27:41 crc kubenswrapper[4998]: I1203 16:27:41.623066 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5" containerName="nova-api-api" containerID="cri-o://872657cb4d2b5e6f6cb1eeab3773bb0cad42991d86cc4d4d3c03d3aeae30eb02" gracePeriod=30 Dec 03 16:27:41 crc kubenswrapper[4998]: I1203 16:27:41.958998 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:27:41 crc kubenswrapper[4998]: I1203 16:27:41.959547 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" containerName="ceilometer-central-agent" containerID="cri-o://4215d5e99baf226e7f679ab78b4f3a7faf72ab49a17d64c14ea966671e5c53c1" gracePeriod=30 Dec 03 16:27:41 crc kubenswrapper[4998]: I1203 16:27:41.959688 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" containerName="proxy-httpd" containerID="cri-o://8b7f798da97884f65fa3da02156f3092bac3a18f0058055f6c7a8c2d74333875" gracePeriod=30 Dec 03 16:27:41 crc kubenswrapper[4998]: I1203 16:27:41.959722 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" containerName="sg-core" containerID="cri-o://44b9c98072d88268c274fdada0200c7d6641a4b6f7548997a7c9b7bc3ef1a976" gracePeriod=30 Dec 03 16:27:41 crc kubenswrapper[4998]: I1203 16:27:41.959734 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" containerName="ceilometer-notification-agent" containerID="cri-o://5866a4f0439acbebc83da557dba8915695913a9bbc0b91e588977170748394dd" gracePeriod=30 Dec 03 16:27:42 crc kubenswrapper[4998]: I1203 16:27:42.444285 4998 generic.go:334] "Generic (PLEG): container finished" podID="fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5" containerID="6f8e7e8f2329d79fc5b615cf264882592e828b8c5cbc63688e35c28d8af30f5a" exitCode=143 Dec 03 16:27:42 crc kubenswrapper[4998]: I1203 16:27:42.444360 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5","Type":"ContainerDied","Data":"6f8e7e8f2329d79fc5b615cf264882592e828b8c5cbc63688e35c28d8af30f5a"} Dec 03 16:27:42 crc kubenswrapper[4998]: I1203 16:27:42.448979 4998 generic.go:334] "Generic (PLEG): container finished" podID="8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" containerID="8b7f798da97884f65fa3da02156f3092bac3a18f0058055f6c7a8c2d74333875" exitCode=0 Dec 03 16:27:42 crc kubenswrapper[4998]: I1203 16:27:42.449027 4998 generic.go:334] "Generic (PLEG): container finished" podID="8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" containerID="44b9c98072d88268c274fdada0200c7d6641a4b6f7548997a7c9b7bc3ef1a976" exitCode=2 Dec 03 16:27:42 crc kubenswrapper[4998]: I1203 16:27:42.449037 4998 generic.go:334] "Generic (PLEG): container finished" podID="8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" containerID="4215d5e99baf226e7f679ab78b4f3a7faf72ab49a17d64c14ea966671e5c53c1" exitCode=0 Dec 03 16:27:42 crc kubenswrapper[4998]: I1203 16:27:42.449124 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af","Type":"ContainerDied","Data":"8b7f798da97884f65fa3da02156f3092bac3a18f0058055f6c7a8c2d74333875"} Dec 03 16:27:42 crc kubenswrapper[4998]: I1203 16:27:42.449173 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af","Type":"ContainerDied","Data":"44b9c98072d88268c274fdada0200c7d6641a4b6f7548997a7c9b7bc3ef1a976"} Dec 03 16:27:42 crc kubenswrapper[4998]: I1203 16:27:42.449191 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af","Type":"ContainerDied","Data":"4215d5e99baf226e7f679ab78b4f3a7faf72ab49a17d64c14ea966671e5c53c1"} Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.477808 4998 generic.go:334] "Generic (PLEG): container finished" podID="8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" containerID="5866a4f0439acbebc83da557dba8915695913a9bbc0b91e588977170748394dd" exitCode=0 Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.478145 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af","Type":"ContainerDied","Data":"5866a4f0439acbebc83da557dba8915695913a9bbc0b91e588977170748394dd"} Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.483630 4998 generic.go:334] "Generic (PLEG): container finished" podID="fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5" containerID="872657cb4d2b5e6f6cb1eeab3773bb0cad42991d86cc4d4d3c03d3aeae30eb02" exitCode=0 Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.483678 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5","Type":"ContainerDied","Data":"872657cb4d2b5e6f6cb1eeab3773bb0cad42991d86cc4d4d3c03d3aeae30eb02"} Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.840014 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.845840 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.870195 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-sg-core-conf-yaml\") pod \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.870343 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-log-httpd\") pod \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.870384 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-ceilometer-tls-certs\") pod \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.871389 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" (UID: "8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.871611 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5-combined-ca-bundle\") pod \"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5\" (UID: \"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5\") " Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.871646 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72qtb\" (UniqueName: \"kubernetes.io/projected/fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5-kube-api-access-72qtb\") pod \"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5\" (UID: \"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5\") " Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.871679 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-config-data\") pod \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.871713 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5-config-data\") pod \"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5\" (UID: \"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5\") " Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.871732 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-combined-ca-bundle\") pod \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.871813 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q42cn\" (UniqueName: \"kubernetes.io/projected/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-kube-api-access-q42cn\") pod \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.871890 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-scripts\") pod \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.871922 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-run-httpd\") pod \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\" (UID: \"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af\") " Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.871993 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5-logs\") pod \"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5\" (UID: \"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5\") " Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.873233 4998 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.873538 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5-logs" (OuterVolumeSpecName: "logs") pod "fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5" (UID: "fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.874783 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" (UID: "8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.881935 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-scripts" (OuterVolumeSpecName: "scripts") pod "8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" (UID: "8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.903962 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-kube-api-access-q42cn" (OuterVolumeSpecName: "kube-api-access-q42cn") pod "8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" (UID: "8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af"). InnerVolumeSpecName "kube-api-access-q42cn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.904069 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5-kube-api-access-72qtb" (OuterVolumeSpecName: "kube-api-access-72qtb") pod "fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5" (UID: "fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5"). InnerVolumeSpecName "kube-api-access-72qtb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.918280 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5-config-data" (OuterVolumeSpecName: "config-data") pod "fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5" (UID: "fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.938501 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" (UID: "8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.967052 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5" (UID: "fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.975891 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.975937 4998 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.975953 4998 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5-logs\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.975965 4998 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.975979 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.975990 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72qtb\" (UniqueName: \"kubernetes.io/projected/fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5-kube-api-access-72qtb\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.976003 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:43 crc kubenswrapper[4998]: I1203 16:27:43.976014 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q42cn\" (UniqueName: \"kubernetes.io/projected/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-kube-api-access-q42cn\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.009436 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" (UID: "8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.027245 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" (UID: "8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.066912 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-config-data" (OuterVolumeSpecName: "config-data") pod "8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" (UID: "8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.078374 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.078425 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.078440 4998 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.498579 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5","Type":"ContainerDied","Data":"fe2476b33790a54d9f75509e1d7bcaafabffc92c7cfc3a64f71fc8224a71dc65"} Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.498647 4998 scope.go:117] "RemoveContainer" containerID="872657cb4d2b5e6f6cb1eeab3773bb0cad42991d86cc4d4d3c03d3aeae30eb02" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.498775 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.508147 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af","Type":"ContainerDied","Data":"7369788119e596e8216be3c0c58d630cf883d6dadbb9b5bbd9084112ca939a85"} Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.508325 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.560027 4998 scope.go:117] "RemoveContainer" containerID="6f8e7e8f2329d79fc5b615cf264882592e828b8c5cbc63688e35c28d8af30f5a" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.642326 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.642656 4998 scope.go:117] "RemoveContainer" containerID="8b7f798da97884f65fa3da02156f3092bac3a18f0058055f6c7a8c2d74333875" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.657479 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.668170 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.673288 4998 scope.go:117] "RemoveContainer" containerID="44b9c98072d88268c274fdada0200c7d6641a4b6f7548997a7c9b7bc3ef1a976" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.678879 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.687400 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 16:27:44 crc kubenswrapper[4998]: E1203 16:27:44.687842 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" containerName="proxy-httpd" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.687858 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" containerName="proxy-httpd" Dec 03 16:27:44 crc kubenswrapper[4998]: E1203 16:27:44.687877 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" containerName="sg-core" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.687884 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" containerName="sg-core" Dec 03 16:27:44 crc kubenswrapper[4998]: E1203 16:27:44.687894 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5" containerName="nova-api-api" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.687901 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5" containerName="nova-api-api" Dec 03 16:27:44 crc kubenswrapper[4998]: E1203 16:27:44.687930 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" containerName="ceilometer-central-agent" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.687937 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" containerName="ceilometer-central-agent" Dec 03 16:27:44 crc kubenswrapper[4998]: E1203 16:27:44.687950 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5" containerName="nova-api-log" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.687955 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5" containerName="nova-api-log" Dec 03 16:27:44 crc kubenswrapper[4998]: E1203 16:27:44.687966 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" containerName="ceilometer-notification-agent" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.687975 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" containerName="ceilometer-notification-agent" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.688148 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" containerName="proxy-httpd" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.688163 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5" containerName="nova-api-log" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.688178 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5" containerName="nova-api-api" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.688186 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" containerName="sg-core" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.688197 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" containerName="ceilometer-central-agent" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.688209 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" containerName="ceilometer-notification-agent" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.689318 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.691980 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.698726 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.701261 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.707666 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.714406 4998 scope.go:117] "RemoveContainer" containerID="5866a4f0439acbebc83da557dba8915695913a9bbc0b91e588977170748394dd" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.721437 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.724821 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.730525 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.730606 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.730664 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.731604 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.754834 4998 scope.go:117] "RemoveContainer" containerID="4215d5e99baf226e7f679ab78b4f3a7faf72ab49a17d64c14ea966671e5c53c1" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.804011 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8849882f-39f2-45f3-86e0-e195e81f003c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8849882f-39f2-45f3-86e0-e195e81f003c\") " pod="openstack/nova-api-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.804077 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa21daa5-d0b4-4d69-a959-f79a5869d9fe-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"fa21daa5-d0b4-4d69-a959-f79a5869d9fe\") " pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.804104 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa21daa5-d0b4-4d69-a959-f79a5869d9fe-log-httpd\") pod \"ceilometer-0\" (UID: \"fa21daa5-d0b4-4d69-a959-f79a5869d9fe\") " pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.804122 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8849882f-39f2-45f3-86e0-e195e81f003c-config-data\") pod \"nova-api-0\" (UID: \"8849882f-39f2-45f3-86e0-e195e81f003c\") " pod="openstack/nova-api-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.804140 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fa21daa5-d0b4-4d69-a959-f79a5869d9fe-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fa21daa5-d0b4-4d69-a959-f79a5869d9fe\") " pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.804183 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa21daa5-d0b4-4d69-a959-f79a5869d9fe-run-httpd\") pod \"ceilometer-0\" (UID: \"fa21daa5-d0b4-4d69-a959-f79a5869d9fe\") " pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.804213 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8849882f-39f2-45f3-86e0-e195e81f003c-public-tls-certs\") pod \"nova-api-0\" (UID: \"8849882f-39f2-45f3-86e0-e195e81f003c\") " pod="openstack/nova-api-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.804283 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa21daa5-d0b4-4d69-a959-f79a5869d9fe-scripts\") pod \"ceilometer-0\" (UID: \"fa21daa5-d0b4-4d69-a959-f79a5869d9fe\") " pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.804316 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8849882f-39f2-45f3-86e0-e195e81f003c-logs\") pod \"nova-api-0\" (UID: \"8849882f-39f2-45f3-86e0-e195e81f003c\") " pod="openstack/nova-api-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.804332 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa21daa5-d0b4-4d69-a959-f79a5869d9fe-config-data\") pod \"ceilometer-0\" (UID: \"fa21daa5-d0b4-4d69-a959-f79a5869d9fe\") " pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.804352 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4jzq\" (UniqueName: \"kubernetes.io/projected/8849882f-39f2-45f3-86e0-e195e81f003c-kube-api-access-z4jzq\") pod \"nova-api-0\" (UID: \"8849882f-39f2-45f3-86e0-e195e81f003c\") " pod="openstack/nova-api-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.804396 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8849882f-39f2-45f3-86e0-e195e81f003c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8849882f-39f2-45f3-86e0-e195e81f003c\") " pod="openstack/nova-api-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.804417 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa21daa5-d0b4-4d69-a959-f79a5869d9fe-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fa21daa5-d0b4-4d69-a959-f79a5869d9fe\") " pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.804439 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngkn9\" (UniqueName: \"kubernetes.io/projected/fa21daa5-d0b4-4d69-a959-f79a5869d9fe-kube-api-access-ngkn9\") pod \"ceilometer-0\" (UID: \"fa21daa5-d0b4-4d69-a959-f79a5869d9fe\") " pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.905962 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa21daa5-d0b4-4d69-a959-f79a5869d9fe-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fa21daa5-d0b4-4d69-a959-f79a5869d9fe\") " pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.906027 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngkn9\" (UniqueName: \"kubernetes.io/projected/fa21daa5-d0b4-4d69-a959-f79a5869d9fe-kube-api-access-ngkn9\") pod \"ceilometer-0\" (UID: \"fa21daa5-d0b4-4d69-a959-f79a5869d9fe\") " pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.906069 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8849882f-39f2-45f3-86e0-e195e81f003c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8849882f-39f2-45f3-86e0-e195e81f003c\") " pod="openstack/nova-api-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.906092 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa21daa5-d0b4-4d69-a959-f79a5869d9fe-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"fa21daa5-d0b4-4d69-a959-f79a5869d9fe\") " pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.906110 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa21daa5-d0b4-4d69-a959-f79a5869d9fe-log-httpd\") pod \"ceilometer-0\" (UID: \"fa21daa5-d0b4-4d69-a959-f79a5869d9fe\") " pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.906127 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8849882f-39f2-45f3-86e0-e195e81f003c-config-data\") pod \"nova-api-0\" (UID: \"8849882f-39f2-45f3-86e0-e195e81f003c\") " pod="openstack/nova-api-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.906144 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fa21daa5-d0b4-4d69-a959-f79a5869d9fe-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fa21daa5-d0b4-4d69-a959-f79a5869d9fe\") " pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.906159 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa21daa5-d0b4-4d69-a959-f79a5869d9fe-run-httpd\") pod \"ceilometer-0\" (UID: \"fa21daa5-d0b4-4d69-a959-f79a5869d9fe\") " pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.906186 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8849882f-39f2-45f3-86e0-e195e81f003c-public-tls-certs\") pod \"nova-api-0\" (UID: \"8849882f-39f2-45f3-86e0-e195e81f003c\") " pod="openstack/nova-api-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.906234 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa21daa5-d0b4-4d69-a959-f79a5869d9fe-scripts\") pod \"ceilometer-0\" (UID: \"fa21daa5-d0b4-4d69-a959-f79a5869d9fe\") " pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.906265 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8849882f-39f2-45f3-86e0-e195e81f003c-logs\") pod \"nova-api-0\" (UID: \"8849882f-39f2-45f3-86e0-e195e81f003c\") " pod="openstack/nova-api-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.906279 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa21daa5-d0b4-4d69-a959-f79a5869d9fe-config-data\") pod \"ceilometer-0\" (UID: \"fa21daa5-d0b4-4d69-a959-f79a5869d9fe\") " pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.906296 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4jzq\" (UniqueName: \"kubernetes.io/projected/8849882f-39f2-45f3-86e0-e195e81f003c-kube-api-access-z4jzq\") pod \"nova-api-0\" (UID: \"8849882f-39f2-45f3-86e0-e195e81f003c\") " pod="openstack/nova-api-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.906338 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8849882f-39f2-45f3-86e0-e195e81f003c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8849882f-39f2-45f3-86e0-e195e81f003c\") " pod="openstack/nova-api-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.906743 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa21daa5-d0b4-4d69-a959-f79a5869d9fe-log-httpd\") pod \"ceilometer-0\" (UID: \"fa21daa5-d0b4-4d69-a959-f79a5869d9fe\") " pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.907044 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fa21daa5-d0b4-4d69-a959-f79a5869d9fe-run-httpd\") pod \"ceilometer-0\" (UID: \"fa21daa5-d0b4-4d69-a959-f79a5869d9fe\") " pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.911129 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8849882f-39f2-45f3-86e0-e195e81f003c-logs\") pod \"nova-api-0\" (UID: \"8849882f-39f2-45f3-86e0-e195e81f003c\") " pod="openstack/nova-api-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.913445 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8849882f-39f2-45f3-86e0-e195e81f003c-public-tls-certs\") pod \"nova-api-0\" (UID: \"8849882f-39f2-45f3-86e0-e195e81f003c\") " pod="openstack/nova-api-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.913768 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa21daa5-d0b4-4d69-a959-f79a5869d9fe-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fa21daa5-d0b4-4d69-a959-f79a5869d9fe\") " pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.929365 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fa21daa5-d0b4-4d69-a959-f79a5869d9fe-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fa21daa5-d0b4-4d69-a959-f79a5869d9fe\") " pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.932421 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8849882f-39f2-45f3-86e0-e195e81f003c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8849882f-39f2-45f3-86e0-e195e81f003c\") " pod="openstack/nova-api-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.936498 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8849882f-39f2-45f3-86e0-e195e81f003c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8849882f-39f2-45f3-86e0-e195e81f003c\") " pod="openstack/nova-api-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.936591 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8849882f-39f2-45f3-86e0-e195e81f003c-config-data\") pod \"nova-api-0\" (UID: \"8849882f-39f2-45f3-86e0-e195e81f003c\") " pod="openstack/nova-api-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.937070 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fa21daa5-d0b4-4d69-a959-f79a5869d9fe-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"fa21daa5-d0b4-4d69-a959-f79a5869d9fe\") " pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.937558 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fa21daa5-d0b4-4d69-a959-f79a5869d9fe-scripts\") pod \"ceilometer-0\" (UID: \"fa21daa5-d0b4-4d69-a959-f79a5869d9fe\") " pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.937611 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa21daa5-d0b4-4d69-a959-f79a5869d9fe-config-data\") pod \"ceilometer-0\" (UID: \"fa21daa5-d0b4-4d69-a959-f79a5869d9fe\") " pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.940926 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngkn9\" (UniqueName: \"kubernetes.io/projected/fa21daa5-d0b4-4d69-a959-f79a5869d9fe-kube-api-access-ngkn9\") pod \"ceilometer-0\" (UID: \"fa21daa5-d0b4-4d69-a959-f79a5869d9fe\") " pod="openstack/ceilometer-0" Dec 03 16:27:44 crc kubenswrapper[4998]: I1203 16:27:44.943255 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4jzq\" (UniqueName: \"kubernetes.io/projected/8849882f-39f2-45f3-86e0-e195e81f003c-kube-api-access-z4jzq\") pod \"nova-api-0\" (UID: \"8849882f-39f2-45f3-86e0-e195e81f003c\") " pod="openstack/nova-api-0" Dec 03 16:27:45 crc kubenswrapper[4998]: I1203 16:27:45.022478 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 16:27:45 crc kubenswrapper[4998]: I1203 16:27:45.058897 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 16:27:45 crc kubenswrapper[4998]: I1203 16:27:45.500723 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 16:27:45 crc kubenswrapper[4998]: I1203 16:27:45.597903 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 16:27:45 crc kubenswrapper[4998]: W1203 16:27:45.602355 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa21daa5_d0b4_4d69_a959_f79a5869d9fe.slice/crio-8146c9179ba887564ba4308ccbc56318f7e63b231d000477fff4245b3447c0b6 WatchSource:0}: Error finding container 8146c9179ba887564ba4308ccbc56318f7e63b231d000477fff4245b3447c0b6: Status 404 returned error can't find the container with id 8146c9179ba887564ba4308ccbc56318f7e63b231d000477fff4245b3447c0b6 Dec 03 16:27:45 crc kubenswrapper[4998]: I1203 16:27:45.691297 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af" path="/var/lib/kubelet/pods/8032c83a-f3f0-4cce-a1f3-7ec9e26ba7af/volumes" Dec 03 16:27:45 crc kubenswrapper[4998]: I1203 16:27:45.693922 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5" path="/var/lib/kubelet/pods/fc0eef29-b63e-40cc-ae2c-0ae5fd5bf3d5/volumes" Dec 03 16:27:45 crc kubenswrapper[4998]: I1203 16:27:45.794812 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:45 crc kubenswrapper[4998]: I1203 16:27:45.814678 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:46 crc kubenswrapper[4998]: I1203 16:27:46.551952 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8849882f-39f2-45f3-86e0-e195e81f003c","Type":"ContainerStarted","Data":"e7ec41fba0cfef6c75ae6dc4be59df4fd8de9d824c1354bd6c58740ec7752f7b"} Dec 03 16:27:46 crc kubenswrapper[4998]: I1203 16:27:46.552583 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8849882f-39f2-45f3-86e0-e195e81f003c","Type":"ContainerStarted","Data":"5239376b905ac4e009c765940d81a02c25f161d572cbcd0be748e0ae5d6fd989"} Dec 03 16:27:46 crc kubenswrapper[4998]: I1203 16:27:46.552605 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8849882f-39f2-45f3-86e0-e195e81f003c","Type":"ContainerStarted","Data":"28328a7dbb6d8ab1374b82ae7c43bbe30cbcaf371e347f2bba9be0cbbd095253"} Dec 03 16:27:46 crc kubenswrapper[4998]: I1203 16:27:46.554647 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fa21daa5-d0b4-4d69-a959-f79a5869d9fe","Type":"ContainerStarted","Data":"e6fcf67814a429b929a44900a64f5e962042c27565b6a6795161f724da524f38"} Dec 03 16:27:46 crc kubenswrapper[4998]: I1203 16:27:46.554688 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fa21daa5-d0b4-4d69-a959-f79a5869d9fe","Type":"ContainerStarted","Data":"0cb14b969a7316e0ea9d01aff7e203e8497639c1805d83fa71ad4ab7e22a855a"} Dec 03 16:27:46 crc kubenswrapper[4998]: I1203 16:27:46.554703 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fa21daa5-d0b4-4d69-a959-f79a5869d9fe","Type":"ContainerStarted","Data":"8146c9179ba887564ba4308ccbc56318f7e63b231d000477fff4245b3447c0b6"} Dec 03 16:27:46 crc kubenswrapper[4998]: I1203 16:27:46.606490 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 03 16:27:46 crc kubenswrapper[4998]: I1203 16:27:46.627602 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.627570939 podStartE2EDuration="2.627570939s" podCreationTimestamp="2025-12-03 16:27:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:27:46.587201369 +0000 UTC m=+1445.198901622" watchObservedRunningTime="2025-12-03 16:27:46.627570939 +0000 UTC m=+1445.239271192" Dec 03 16:27:46 crc kubenswrapper[4998]: I1203 16:27:46.776412 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-dlfdq"] Dec 03 16:27:46 crc kubenswrapper[4998]: I1203 16:27:46.778576 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dlfdq" Dec 03 16:27:46 crc kubenswrapper[4998]: I1203 16:27:46.788944 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 03 16:27:46 crc kubenswrapper[4998]: I1203 16:27:46.789254 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 03 16:27:46 crc kubenswrapper[4998]: I1203 16:27:46.826867 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-dlfdq"] Dec 03 16:27:46 crc kubenswrapper[4998]: I1203 16:27:46.843680 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53aaa2e5-6add-4615-9aa1-9af99d56ea04-scripts\") pod \"nova-cell1-cell-mapping-dlfdq\" (UID: \"53aaa2e5-6add-4615-9aa1-9af99d56ea04\") " pod="openstack/nova-cell1-cell-mapping-dlfdq" Dec 03 16:27:46 crc kubenswrapper[4998]: I1203 16:27:46.843741 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtm8p\" (UniqueName: \"kubernetes.io/projected/53aaa2e5-6add-4615-9aa1-9af99d56ea04-kube-api-access-wtm8p\") pod \"nova-cell1-cell-mapping-dlfdq\" (UID: \"53aaa2e5-6add-4615-9aa1-9af99d56ea04\") " pod="openstack/nova-cell1-cell-mapping-dlfdq" Dec 03 16:27:46 crc kubenswrapper[4998]: I1203 16:27:46.843800 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53aaa2e5-6add-4615-9aa1-9af99d56ea04-config-data\") pod \"nova-cell1-cell-mapping-dlfdq\" (UID: \"53aaa2e5-6add-4615-9aa1-9af99d56ea04\") " pod="openstack/nova-cell1-cell-mapping-dlfdq" Dec 03 16:27:46 crc kubenswrapper[4998]: I1203 16:27:46.843858 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53aaa2e5-6add-4615-9aa1-9af99d56ea04-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dlfdq\" (UID: \"53aaa2e5-6add-4615-9aa1-9af99d56ea04\") " pod="openstack/nova-cell1-cell-mapping-dlfdq" Dec 03 16:27:46 crc kubenswrapper[4998]: I1203 16:27:46.945196 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53aaa2e5-6add-4615-9aa1-9af99d56ea04-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dlfdq\" (UID: \"53aaa2e5-6add-4615-9aa1-9af99d56ea04\") " pod="openstack/nova-cell1-cell-mapping-dlfdq" Dec 03 16:27:46 crc kubenswrapper[4998]: I1203 16:27:46.945686 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53aaa2e5-6add-4615-9aa1-9af99d56ea04-scripts\") pod \"nova-cell1-cell-mapping-dlfdq\" (UID: \"53aaa2e5-6add-4615-9aa1-9af99d56ea04\") " pod="openstack/nova-cell1-cell-mapping-dlfdq" Dec 03 16:27:46 crc kubenswrapper[4998]: I1203 16:27:46.945713 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtm8p\" (UniqueName: \"kubernetes.io/projected/53aaa2e5-6add-4615-9aa1-9af99d56ea04-kube-api-access-wtm8p\") pod \"nova-cell1-cell-mapping-dlfdq\" (UID: \"53aaa2e5-6add-4615-9aa1-9af99d56ea04\") " pod="openstack/nova-cell1-cell-mapping-dlfdq" Dec 03 16:27:46 crc kubenswrapper[4998]: I1203 16:27:46.945767 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53aaa2e5-6add-4615-9aa1-9af99d56ea04-config-data\") pod \"nova-cell1-cell-mapping-dlfdq\" (UID: \"53aaa2e5-6add-4615-9aa1-9af99d56ea04\") " pod="openstack/nova-cell1-cell-mapping-dlfdq" Dec 03 16:27:46 crc kubenswrapper[4998]: I1203 16:27:46.952071 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53aaa2e5-6add-4615-9aa1-9af99d56ea04-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dlfdq\" (UID: \"53aaa2e5-6add-4615-9aa1-9af99d56ea04\") " pod="openstack/nova-cell1-cell-mapping-dlfdq" Dec 03 16:27:46 crc kubenswrapper[4998]: I1203 16:27:46.956218 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53aaa2e5-6add-4615-9aa1-9af99d56ea04-scripts\") pod \"nova-cell1-cell-mapping-dlfdq\" (UID: \"53aaa2e5-6add-4615-9aa1-9af99d56ea04\") " pod="openstack/nova-cell1-cell-mapping-dlfdq" Dec 03 16:27:46 crc kubenswrapper[4998]: I1203 16:27:46.956736 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53aaa2e5-6add-4615-9aa1-9af99d56ea04-config-data\") pod \"nova-cell1-cell-mapping-dlfdq\" (UID: \"53aaa2e5-6add-4615-9aa1-9af99d56ea04\") " pod="openstack/nova-cell1-cell-mapping-dlfdq" Dec 03 16:27:46 crc kubenswrapper[4998]: I1203 16:27:46.966623 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtm8p\" (UniqueName: \"kubernetes.io/projected/53aaa2e5-6add-4615-9aa1-9af99d56ea04-kube-api-access-wtm8p\") pod \"nova-cell1-cell-mapping-dlfdq\" (UID: \"53aaa2e5-6add-4615-9aa1-9af99d56ea04\") " pod="openstack/nova-cell1-cell-mapping-dlfdq" Dec 03 16:27:47 crc kubenswrapper[4998]: I1203 16:27:47.129853 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dlfdq" Dec 03 16:27:47 crc kubenswrapper[4998]: I1203 16:27:47.571597 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fa21daa5-d0b4-4d69-a959-f79a5869d9fe","Type":"ContainerStarted","Data":"821cdeb56c3954b599ab45e717e8f10ce1791f35f21f8d7db579e7e910384d1e"} Dec 03 16:27:47 crc kubenswrapper[4998]: I1203 16:27:47.655918 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-dlfdq"] Dec 03 16:27:48 crc kubenswrapper[4998]: I1203 16:27:48.580803 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dlfdq" event={"ID":"53aaa2e5-6add-4615-9aa1-9af99d56ea04","Type":"ContainerStarted","Data":"d7a9fa9786f0425c5472b4b8e59b0cb7bf46e79182787c3b5551ee7e4116bc06"} Dec 03 16:27:48 crc kubenswrapper[4998]: I1203 16:27:48.581517 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dlfdq" event={"ID":"53aaa2e5-6add-4615-9aa1-9af99d56ea04","Type":"ContainerStarted","Data":"07bdb08793bc58799f24c7a21cb3011c7192fc07b03e87ad2772b0464881e716"} Dec 03 16:27:48 crc kubenswrapper[4998]: I1203 16:27:48.585494 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fa21daa5-d0b4-4d69-a959-f79a5869d9fe","Type":"ContainerStarted","Data":"456adf1340747ffbf73e8fd51a077c57e9209958e9832967a6eb12813ebdc4b0"} Dec 03 16:27:48 crc kubenswrapper[4998]: I1203 16:27:48.586386 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 16:27:48 crc kubenswrapper[4998]: I1203 16:27:48.620134 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-dlfdq" podStartSLOduration=2.620115216 podStartE2EDuration="2.620115216s" podCreationTimestamp="2025-12-03 16:27:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:27:48.603791381 +0000 UTC m=+1447.215491594" watchObservedRunningTime="2025-12-03 16:27:48.620115216 +0000 UTC m=+1447.231815439" Dec 03 16:27:48 crc kubenswrapper[4998]: I1203 16:27:48.634889 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.05338113 podStartE2EDuration="4.634863151s" podCreationTimestamp="2025-12-03 16:27:44 +0000 UTC" firstStartedPulling="2025-12-03 16:27:45.60545096 +0000 UTC m=+1444.217151183" lastFinishedPulling="2025-12-03 16:27:48.186932981 +0000 UTC m=+1446.798633204" observedRunningTime="2025-12-03 16:27:48.62592286 +0000 UTC m=+1447.237623113" watchObservedRunningTime="2025-12-03 16:27:48.634863151 +0000 UTC m=+1447.246563394" Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.118216 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.186717 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7995555d47-42cp9"] Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.187049 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7995555d47-42cp9" podUID="c435b8fa-d8c6-49fb-b339-edeb6a3f182c" containerName="dnsmasq-dns" containerID="cri-o://c6719e5a27b5be3a5194c080d34db81c4c39ac724df7ddabfa3d9de6ce711dc9" gracePeriod=10 Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.646604 4998 generic.go:334] "Generic (PLEG): container finished" podID="c435b8fa-d8c6-49fb-b339-edeb6a3f182c" containerID="c6719e5a27b5be3a5194c080d34db81c4c39ac724df7ddabfa3d9de6ce711dc9" exitCode=0 Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.647768 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7995555d47-42cp9" event={"ID":"c435b8fa-d8c6-49fb-b339-edeb6a3f182c","Type":"ContainerDied","Data":"c6719e5a27b5be3a5194c080d34db81c4c39ac724df7ddabfa3d9de6ce711dc9"} Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.647897 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7995555d47-42cp9" event={"ID":"c435b8fa-d8c6-49fb-b339-edeb6a3f182c","Type":"ContainerDied","Data":"1b71958fe17dab436da10becd2e5433e45b3b888e7dd2bcac07ecfcfb8a28e51"} Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.647953 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b71958fe17dab436da10becd2e5433e45b3b888e7dd2bcac07ecfcfb8a28e51" Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.695810 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7995555d47-42cp9" Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.736553 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-dns-swift-storage-0\") pod \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\" (UID: \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\") " Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.736642 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhpck\" (UniqueName: \"kubernetes.io/projected/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-kube-api-access-vhpck\") pod \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\" (UID: \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\") " Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.736714 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-dns-svc\") pod \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\" (UID: \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\") " Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.736823 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-ovsdbserver-sb\") pod \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\" (UID: \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\") " Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.736844 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-config\") pod \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\" (UID: \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\") " Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.736878 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-ovsdbserver-nb\") pod \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\" (UID: \"c435b8fa-d8c6-49fb-b339-edeb6a3f182c\") " Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.762482 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-kube-api-access-vhpck" (OuterVolumeSpecName: "kube-api-access-vhpck") pod "c435b8fa-d8c6-49fb-b339-edeb6a3f182c" (UID: "c435b8fa-d8c6-49fb-b339-edeb6a3f182c"). InnerVolumeSpecName "kube-api-access-vhpck". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.799474 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c435b8fa-d8c6-49fb-b339-edeb6a3f182c" (UID: "c435b8fa-d8c6-49fb-b339-edeb6a3f182c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.804268 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-config" (OuterVolumeSpecName: "config") pod "c435b8fa-d8c6-49fb-b339-edeb6a3f182c" (UID: "c435b8fa-d8c6-49fb-b339-edeb6a3f182c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.827258 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c435b8fa-d8c6-49fb-b339-edeb6a3f182c" (UID: "c435b8fa-d8c6-49fb-b339-edeb6a3f182c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.827304 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c435b8fa-d8c6-49fb-b339-edeb6a3f182c" (UID: "c435b8fa-d8c6-49fb-b339-edeb6a3f182c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.840167 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhpck\" (UniqueName: \"kubernetes.io/projected/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-kube-api-access-vhpck\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.840250 4998 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.840268 4998 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.840281 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.840295 4998 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.843400 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c435b8fa-d8c6-49fb-b339-edeb6a3f182c" (UID: "c435b8fa-d8c6-49fb-b339-edeb6a3f182c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:27:49 crc kubenswrapper[4998]: I1203 16:27:49.941773 4998 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c435b8fa-d8c6-49fb-b339-edeb6a3f182c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:50 crc kubenswrapper[4998]: I1203 16:27:50.656896 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7995555d47-42cp9" Dec 03 16:27:50 crc kubenswrapper[4998]: I1203 16:27:50.694320 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7995555d47-42cp9"] Dec 03 16:27:50 crc kubenswrapper[4998]: I1203 16:27:50.704556 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7995555d47-42cp9"] Dec 03 16:27:51 crc kubenswrapper[4998]: I1203 16:27:51.691520 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c435b8fa-d8c6-49fb-b339-edeb6a3f182c" path="/var/lib/kubelet/pods/c435b8fa-d8c6-49fb-b339-edeb6a3f182c/volumes" Dec 03 16:27:53 crc kubenswrapper[4998]: I1203 16:27:53.692016 4998 generic.go:334] "Generic (PLEG): container finished" podID="53aaa2e5-6add-4615-9aa1-9af99d56ea04" containerID="d7a9fa9786f0425c5472b4b8e59b0cb7bf46e79182787c3b5551ee7e4116bc06" exitCode=0 Dec 03 16:27:53 crc kubenswrapper[4998]: I1203 16:27:53.704983 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dlfdq" event={"ID":"53aaa2e5-6add-4615-9aa1-9af99d56ea04","Type":"ContainerDied","Data":"d7a9fa9786f0425c5472b4b8e59b0cb7bf46e79182787c3b5551ee7e4116bc06"} Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.023611 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.024030 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.168055 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dlfdq" Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.267315 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53aaa2e5-6add-4615-9aa1-9af99d56ea04-scripts\") pod \"53aaa2e5-6add-4615-9aa1-9af99d56ea04\" (UID: \"53aaa2e5-6add-4615-9aa1-9af99d56ea04\") " Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.268119 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtm8p\" (UniqueName: \"kubernetes.io/projected/53aaa2e5-6add-4615-9aa1-9af99d56ea04-kube-api-access-wtm8p\") pod \"53aaa2e5-6add-4615-9aa1-9af99d56ea04\" (UID: \"53aaa2e5-6add-4615-9aa1-9af99d56ea04\") " Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.268446 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53aaa2e5-6add-4615-9aa1-9af99d56ea04-combined-ca-bundle\") pod \"53aaa2e5-6add-4615-9aa1-9af99d56ea04\" (UID: \"53aaa2e5-6add-4615-9aa1-9af99d56ea04\") " Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.268643 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53aaa2e5-6add-4615-9aa1-9af99d56ea04-config-data\") pod \"53aaa2e5-6add-4615-9aa1-9af99d56ea04\" (UID: \"53aaa2e5-6add-4615-9aa1-9af99d56ea04\") " Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.274249 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53aaa2e5-6add-4615-9aa1-9af99d56ea04-scripts" (OuterVolumeSpecName: "scripts") pod "53aaa2e5-6add-4615-9aa1-9af99d56ea04" (UID: "53aaa2e5-6add-4615-9aa1-9af99d56ea04"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.274718 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53aaa2e5-6add-4615-9aa1-9af99d56ea04-kube-api-access-wtm8p" (OuterVolumeSpecName: "kube-api-access-wtm8p") pod "53aaa2e5-6add-4615-9aa1-9af99d56ea04" (UID: "53aaa2e5-6add-4615-9aa1-9af99d56ea04"). InnerVolumeSpecName "kube-api-access-wtm8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.302386 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53aaa2e5-6add-4615-9aa1-9af99d56ea04-config-data" (OuterVolumeSpecName: "config-data") pod "53aaa2e5-6add-4615-9aa1-9af99d56ea04" (UID: "53aaa2e5-6add-4615-9aa1-9af99d56ea04"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.318741 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53aaa2e5-6add-4615-9aa1-9af99d56ea04-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53aaa2e5-6add-4615-9aa1-9af99d56ea04" (UID: "53aaa2e5-6add-4615-9aa1-9af99d56ea04"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.372020 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtm8p\" (UniqueName: \"kubernetes.io/projected/53aaa2e5-6add-4615-9aa1-9af99d56ea04-kube-api-access-wtm8p\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.372066 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53aaa2e5-6add-4615-9aa1-9af99d56ea04-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.372080 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53aaa2e5-6add-4615-9aa1-9af99d56ea04-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.372091 4998 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53aaa2e5-6add-4615-9aa1-9af99d56ea04-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.720429 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dlfdq" event={"ID":"53aaa2e5-6add-4615-9aa1-9af99d56ea04","Type":"ContainerDied","Data":"07bdb08793bc58799f24c7a21cb3011c7192fc07b03e87ad2772b0464881e716"} Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.720738 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07bdb08793bc58799f24c7a21cb3011c7192fc07b03e87ad2772b0464881e716" Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.720472 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dlfdq" Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.923658 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.923988 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="2b29e862-8b36-45b1-a23a-534083def903" containerName="nova-scheduler-scheduler" containerID="cri-o://f7b11db1f404c817ddc5d1b94ead20d1cc0cac2133efa6d55366db7bf306214f" gracePeriod=30 Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.939119 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.939336 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8849882f-39f2-45f3-86e0-e195e81f003c" containerName="nova-api-log" containerID="cri-o://5239376b905ac4e009c765940d81a02c25f161d572cbcd0be748e0ae5d6fd989" gracePeriod=30 Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.939439 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8849882f-39f2-45f3-86e0-e195e81f003c" containerName="nova-api-api" containerID="cri-o://e7ec41fba0cfef6c75ae6dc4be59df4fd8de9d824c1354bd6c58740ec7752f7b" gracePeriod=30 Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.964634 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8849882f-39f2-45f3-86e0-e195e81f003c" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.222:8774/\": EOF" Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.964737 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8849882f-39f2-45f3-86e0-e195e81f003c" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.222:8774/\": EOF" Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.969380 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.969622 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="cd010875-398a-48fc-8fa5-569126c4f67f" containerName="nova-metadata-log" containerID="cri-o://3598305bb163d8ec273d76c345836c5e48807e343bbabc183e9bb5c7b73b3df8" gracePeriod=30 Dec 03 16:27:55 crc kubenswrapper[4998]: I1203 16:27:55.969738 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="cd010875-398a-48fc-8fa5-569126c4f67f" containerName="nova-metadata-metadata" containerID="cri-o://b276c4c9c27b1341e38931073a93f1259bd29c8f45fc78cfe649b567c556b32c" gracePeriod=30 Dec 03 16:27:56 crc kubenswrapper[4998]: I1203 16:27:56.731366 4998 generic.go:334] "Generic (PLEG): container finished" podID="cd010875-398a-48fc-8fa5-569126c4f67f" containerID="3598305bb163d8ec273d76c345836c5e48807e343bbabc183e9bb5c7b73b3df8" exitCode=143 Dec 03 16:27:56 crc kubenswrapper[4998]: I1203 16:27:56.731444 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cd010875-398a-48fc-8fa5-569126c4f67f","Type":"ContainerDied","Data":"3598305bb163d8ec273d76c345836c5e48807e343bbabc183e9bb5c7b73b3df8"} Dec 03 16:27:56 crc kubenswrapper[4998]: I1203 16:27:56.733390 4998 generic.go:334] "Generic (PLEG): container finished" podID="8849882f-39f2-45f3-86e0-e195e81f003c" containerID="5239376b905ac4e009c765940d81a02c25f161d572cbcd0be748e0ae5d6fd989" exitCode=143 Dec 03 16:27:56 crc kubenswrapper[4998]: I1203 16:27:56.733413 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8849882f-39f2-45f3-86e0-e195e81f003c","Type":"ContainerDied","Data":"5239376b905ac4e009c765940d81a02c25f161d572cbcd0be748e0ae5d6fd989"} Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.111225 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.111488 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.302918 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.414278 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd010875-398a-48fc-8fa5-569126c4f67f-nova-metadata-tls-certs\") pod \"cd010875-398a-48fc-8fa5-569126c4f67f\" (UID: \"cd010875-398a-48fc-8fa5-569126c4f67f\") " Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.414422 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd010875-398a-48fc-8fa5-569126c4f67f-logs\") pod \"cd010875-398a-48fc-8fa5-569126c4f67f\" (UID: \"cd010875-398a-48fc-8fa5-569126c4f67f\") " Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.414469 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd010875-398a-48fc-8fa5-569126c4f67f-combined-ca-bundle\") pod \"cd010875-398a-48fc-8fa5-569126c4f67f\" (UID: \"cd010875-398a-48fc-8fa5-569126c4f67f\") " Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.414494 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd010875-398a-48fc-8fa5-569126c4f67f-config-data\") pod \"cd010875-398a-48fc-8fa5-569126c4f67f\" (UID: \"cd010875-398a-48fc-8fa5-569126c4f67f\") " Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.414618 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szwsj\" (UniqueName: \"kubernetes.io/projected/cd010875-398a-48fc-8fa5-569126c4f67f-kube-api-access-szwsj\") pod \"cd010875-398a-48fc-8fa5-569126c4f67f\" (UID: \"cd010875-398a-48fc-8fa5-569126c4f67f\") " Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.417218 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd010875-398a-48fc-8fa5-569126c4f67f-logs" (OuterVolumeSpecName: "logs") pod "cd010875-398a-48fc-8fa5-569126c4f67f" (UID: "cd010875-398a-48fc-8fa5-569126c4f67f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.427975 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd010875-398a-48fc-8fa5-569126c4f67f-kube-api-access-szwsj" (OuterVolumeSpecName: "kube-api-access-szwsj") pod "cd010875-398a-48fc-8fa5-569126c4f67f" (UID: "cd010875-398a-48fc-8fa5-569126c4f67f"). InnerVolumeSpecName "kube-api-access-szwsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.455652 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd010875-398a-48fc-8fa5-569126c4f67f-config-data" (OuterVolumeSpecName: "config-data") pod "cd010875-398a-48fc-8fa5-569126c4f67f" (UID: "cd010875-398a-48fc-8fa5-569126c4f67f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.455930 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd010875-398a-48fc-8fa5-569126c4f67f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd010875-398a-48fc-8fa5-569126c4f67f" (UID: "cd010875-398a-48fc-8fa5-569126c4f67f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.487646 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd010875-398a-48fc-8fa5-569126c4f67f-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "cd010875-398a-48fc-8fa5-569126c4f67f" (UID: "cd010875-398a-48fc-8fa5-569126c4f67f"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.516803 4998 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd010875-398a-48fc-8fa5-569126c4f67f-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.516830 4998 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd010875-398a-48fc-8fa5-569126c4f67f-logs\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.516841 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd010875-398a-48fc-8fa5-569126c4f67f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.516849 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd010875-398a-48fc-8fa5-569126c4f67f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.516859 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szwsj\" (UniqueName: \"kubernetes.io/projected/cd010875-398a-48fc-8fa5-569126c4f67f-kube-api-access-szwsj\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.746005 4998 generic.go:334] "Generic (PLEG): container finished" podID="cd010875-398a-48fc-8fa5-569126c4f67f" containerID="b276c4c9c27b1341e38931073a93f1259bd29c8f45fc78cfe649b567c556b32c" exitCode=0 Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.746267 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cd010875-398a-48fc-8fa5-569126c4f67f","Type":"ContainerDied","Data":"b276c4c9c27b1341e38931073a93f1259bd29c8f45fc78cfe649b567c556b32c"} Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.746290 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cd010875-398a-48fc-8fa5-569126c4f67f","Type":"ContainerDied","Data":"bbc157a93f57519359c12ddddcafb73d83738537de2687889c41da4eee74a7e0"} Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.746314 4998 scope.go:117] "RemoveContainer" containerID="b276c4c9c27b1341e38931073a93f1259bd29c8f45fc78cfe649b567c556b32c" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.746449 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.833354 4998 scope.go:117] "RemoveContainer" containerID="3598305bb163d8ec273d76c345836c5e48807e343bbabc183e9bb5c7b73b3df8" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.857845 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.960914 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.969200 4998 scope.go:117] "RemoveContainer" containerID="b276c4c9c27b1341e38931073a93f1259bd29c8f45fc78cfe649b567c556b32c" Dec 03 16:27:57 crc kubenswrapper[4998]: E1203 16:27:57.972882 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b276c4c9c27b1341e38931073a93f1259bd29c8f45fc78cfe649b567c556b32c\": container with ID starting with b276c4c9c27b1341e38931073a93f1259bd29c8f45fc78cfe649b567c556b32c not found: ID does not exist" containerID="b276c4c9c27b1341e38931073a93f1259bd29c8f45fc78cfe649b567c556b32c" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.972927 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b276c4c9c27b1341e38931073a93f1259bd29c8f45fc78cfe649b567c556b32c"} err="failed to get container status \"b276c4c9c27b1341e38931073a93f1259bd29c8f45fc78cfe649b567c556b32c\": rpc error: code = NotFound desc = could not find container \"b276c4c9c27b1341e38931073a93f1259bd29c8f45fc78cfe649b567c556b32c\": container with ID starting with b276c4c9c27b1341e38931073a93f1259bd29c8f45fc78cfe649b567c556b32c not found: ID does not exist" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.972953 4998 scope.go:117] "RemoveContainer" containerID="3598305bb163d8ec273d76c345836c5e48807e343bbabc183e9bb5c7b73b3df8" Dec 03 16:27:57 crc kubenswrapper[4998]: E1203 16:27:57.992918 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3598305bb163d8ec273d76c345836c5e48807e343bbabc183e9bb5c7b73b3df8\": container with ID starting with 3598305bb163d8ec273d76c345836c5e48807e343bbabc183e9bb5c7b73b3df8 not found: ID does not exist" containerID="3598305bb163d8ec273d76c345836c5e48807e343bbabc183e9bb5c7b73b3df8" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.993182 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3598305bb163d8ec273d76c345836c5e48807e343bbabc183e9bb5c7b73b3df8"} err="failed to get container status \"3598305bb163d8ec273d76c345836c5e48807e343bbabc183e9bb5c7b73b3df8\": rpc error: code = NotFound desc = could not find container \"3598305bb163d8ec273d76c345836c5e48807e343bbabc183e9bb5c7b73b3df8\": container with ID starting with 3598305bb163d8ec273d76c345836c5e48807e343bbabc183e9bb5c7b73b3df8 not found: ID does not exist" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.998836 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 16:27:57 crc kubenswrapper[4998]: E1203 16:27:57.999289 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd010875-398a-48fc-8fa5-569126c4f67f" containerName="nova-metadata-metadata" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.999306 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd010875-398a-48fc-8fa5-569126c4f67f" containerName="nova-metadata-metadata" Dec 03 16:27:57 crc kubenswrapper[4998]: E1203 16:27:57.999323 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c435b8fa-d8c6-49fb-b339-edeb6a3f182c" containerName="dnsmasq-dns" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.999330 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="c435b8fa-d8c6-49fb-b339-edeb6a3f182c" containerName="dnsmasq-dns" Dec 03 16:27:57 crc kubenswrapper[4998]: E1203 16:27:57.999345 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53aaa2e5-6add-4615-9aa1-9af99d56ea04" containerName="nova-manage" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.999352 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="53aaa2e5-6add-4615-9aa1-9af99d56ea04" containerName="nova-manage" Dec 03 16:27:57 crc kubenswrapper[4998]: E1203 16:27:57.999361 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c435b8fa-d8c6-49fb-b339-edeb6a3f182c" containerName="init" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.999366 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="c435b8fa-d8c6-49fb-b339-edeb6a3f182c" containerName="init" Dec 03 16:27:57 crc kubenswrapper[4998]: E1203 16:27:57.999376 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd010875-398a-48fc-8fa5-569126c4f67f" containerName="nova-metadata-log" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.999382 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd010875-398a-48fc-8fa5-569126c4f67f" containerName="nova-metadata-log" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.999550 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd010875-398a-48fc-8fa5-569126c4f67f" containerName="nova-metadata-log" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.999563 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="c435b8fa-d8c6-49fb-b339-edeb6a3f182c" containerName="dnsmasq-dns" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.999576 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd010875-398a-48fc-8fa5-569126c4f67f" containerName="nova-metadata-metadata" Dec 03 16:27:57 crc kubenswrapper[4998]: I1203 16:27:57.999584 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="53aaa2e5-6add-4615-9aa1-9af99d56ea04" containerName="nova-manage" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.000699 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.020268 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.020497 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.038513 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.136793 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff8e487a-c00a-47a8-997f-6a0376aa873d-logs\") pod \"nova-metadata-0\" (UID: \"ff8e487a-c00a-47a8-997f-6a0376aa873d\") " pod="openstack/nova-metadata-0" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.136849 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmxqt\" (UniqueName: \"kubernetes.io/projected/ff8e487a-c00a-47a8-997f-6a0376aa873d-kube-api-access-mmxqt\") pod \"nova-metadata-0\" (UID: \"ff8e487a-c00a-47a8-997f-6a0376aa873d\") " pod="openstack/nova-metadata-0" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.136881 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff8e487a-c00a-47a8-997f-6a0376aa873d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ff8e487a-c00a-47a8-997f-6a0376aa873d\") " pod="openstack/nova-metadata-0" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.136933 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff8e487a-c00a-47a8-997f-6a0376aa873d-config-data\") pod \"nova-metadata-0\" (UID: \"ff8e487a-c00a-47a8-997f-6a0376aa873d\") " pod="openstack/nova-metadata-0" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.136967 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff8e487a-c00a-47a8-997f-6a0376aa873d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ff8e487a-c00a-47a8-997f-6a0376aa873d\") " pod="openstack/nova-metadata-0" Dec 03 16:27:58 crc kubenswrapper[4998]: E1203 16:27:58.215762 4998 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b29e862_8b36_45b1_a23a_534083def903.slice/crio-conmon-f7b11db1f404c817ddc5d1b94ead20d1cc0cac2133efa6d55366db7bf306214f.scope\": RecentStats: unable to find data in memory cache]" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.242853 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff8e487a-c00a-47a8-997f-6a0376aa873d-config-data\") pod \"nova-metadata-0\" (UID: \"ff8e487a-c00a-47a8-997f-6a0376aa873d\") " pod="openstack/nova-metadata-0" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.242914 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff8e487a-c00a-47a8-997f-6a0376aa873d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ff8e487a-c00a-47a8-997f-6a0376aa873d\") " pod="openstack/nova-metadata-0" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.242984 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff8e487a-c00a-47a8-997f-6a0376aa873d-logs\") pod \"nova-metadata-0\" (UID: \"ff8e487a-c00a-47a8-997f-6a0376aa873d\") " pod="openstack/nova-metadata-0" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.243017 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmxqt\" (UniqueName: \"kubernetes.io/projected/ff8e487a-c00a-47a8-997f-6a0376aa873d-kube-api-access-mmxqt\") pod \"nova-metadata-0\" (UID: \"ff8e487a-c00a-47a8-997f-6a0376aa873d\") " pod="openstack/nova-metadata-0" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.243042 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff8e487a-c00a-47a8-997f-6a0376aa873d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ff8e487a-c00a-47a8-997f-6a0376aa873d\") " pod="openstack/nova-metadata-0" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.244385 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff8e487a-c00a-47a8-997f-6a0376aa873d-logs\") pod \"nova-metadata-0\" (UID: \"ff8e487a-c00a-47a8-997f-6a0376aa873d\") " pod="openstack/nova-metadata-0" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.249026 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff8e487a-c00a-47a8-997f-6a0376aa873d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ff8e487a-c00a-47a8-997f-6a0376aa873d\") " pod="openstack/nova-metadata-0" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.249993 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff8e487a-c00a-47a8-997f-6a0376aa873d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ff8e487a-c00a-47a8-997f-6a0376aa873d\") " pod="openstack/nova-metadata-0" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.250420 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff8e487a-c00a-47a8-997f-6a0376aa873d-config-data\") pod \"nova-metadata-0\" (UID: \"ff8e487a-c00a-47a8-997f-6a0376aa873d\") " pod="openstack/nova-metadata-0" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.278354 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmxqt\" (UniqueName: \"kubernetes.io/projected/ff8e487a-c00a-47a8-997f-6a0376aa873d-kube-api-access-mmxqt\") pod \"nova-metadata-0\" (UID: \"ff8e487a-c00a-47a8-997f-6a0376aa873d\") " pod="openstack/nova-metadata-0" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.335712 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.347419 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.459334 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b29e862-8b36-45b1-a23a-534083def903-combined-ca-bundle\") pod \"2b29e862-8b36-45b1-a23a-534083def903\" (UID: \"2b29e862-8b36-45b1-a23a-534083def903\") " Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.459404 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlv55\" (UniqueName: \"kubernetes.io/projected/2b29e862-8b36-45b1-a23a-534083def903-kube-api-access-mlv55\") pod \"2b29e862-8b36-45b1-a23a-534083def903\" (UID: \"2b29e862-8b36-45b1-a23a-534083def903\") " Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.459521 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b29e862-8b36-45b1-a23a-534083def903-config-data\") pod \"2b29e862-8b36-45b1-a23a-534083def903\" (UID: \"2b29e862-8b36-45b1-a23a-534083def903\") " Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.464713 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b29e862-8b36-45b1-a23a-534083def903-kube-api-access-mlv55" (OuterVolumeSpecName: "kube-api-access-mlv55") pod "2b29e862-8b36-45b1-a23a-534083def903" (UID: "2b29e862-8b36-45b1-a23a-534083def903"). InnerVolumeSpecName "kube-api-access-mlv55". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.492046 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b29e862-8b36-45b1-a23a-534083def903-config-data" (OuterVolumeSpecName: "config-data") pod "2b29e862-8b36-45b1-a23a-534083def903" (UID: "2b29e862-8b36-45b1-a23a-534083def903"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.500286 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b29e862-8b36-45b1-a23a-534083def903-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2b29e862-8b36-45b1-a23a-534083def903" (UID: "2b29e862-8b36-45b1-a23a-534083def903"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.561117 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b29e862-8b36-45b1-a23a-534083def903-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.561350 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b29e862-8b36-45b1-a23a-534083def903-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.561362 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlv55\" (UniqueName: \"kubernetes.io/projected/2b29e862-8b36-45b1-a23a-534083def903-kube-api-access-mlv55\") on node \"crc\" DevicePath \"\"" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.759035 4998 generic.go:334] "Generic (PLEG): container finished" podID="2b29e862-8b36-45b1-a23a-534083def903" containerID="f7b11db1f404c817ddc5d1b94ead20d1cc0cac2133efa6d55366db7bf306214f" exitCode=0 Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.759099 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2b29e862-8b36-45b1-a23a-534083def903","Type":"ContainerDied","Data":"f7b11db1f404c817ddc5d1b94ead20d1cc0cac2133efa6d55366db7bf306214f"} Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.759105 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.759126 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2b29e862-8b36-45b1-a23a-534083def903","Type":"ContainerDied","Data":"be2e6757ade78923f7147fcf6c330255631097e98a7ecb255e15dce22120b142"} Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.759142 4998 scope.go:117] "RemoveContainer" containerID="f7b11db1f404c817ddc5d1b94ead20d1cc0cac2133efa6d55366db7bf306214f" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.792874 4998 scope.go:117] "RemoveContainer" containerID="f7b11db1f404c817ddc5d1b94ead20d1cc0cac2133efa6d55366db7bf306214f" Dec 03 16:27:58 crc kubenswrapper[4998]: E1203 16:27:58.793458 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7b11db1f404c817ddc5d1b94ead20d1cc0cac2133efa6d55366db7bf306214f\": container with ID starting with f7b11db1f404c817ddc5d1b94ead20d1cc0cac2133efa6d55366db7bf306214f not found: ID does not exist" containerID="f7b11db1f404c817ddc5d1b94ead20d1cc0cac2133efa6d55366db7bf306214f" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.793519 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7b11db1f404c817ddc5d1b94ead20d1cc0cac2133efa6d55366db7bf306214f"} err="failed to get container status \"f7b11db1f404c817ddc5d1b94ead20d1cc0cac2133efa6d55366db7bf306214f\": rpc error: code = NotFound desc = could not find container \"f7b11db1f404c817ddc5d1b94ead20d1cc0cac2133efa6d55366db7bf306214f\": container with ID starting with f7b11db1f404c817ddc5d1b94ead20d1cc0cac2133efa6d55366db7bf306214f not found: ID does not exist" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.812109 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.831082 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.842120 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 16:27:58 crc kubenswrapper[4998]: E1203 16:27:58.842591 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b29e862-8b36-45b1-a23a-534083def903" containerName="nova-scheduler-scheduler" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.842607 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b29e862-8b36-45b1-a23a-534083def903" containerName="nova-scheduler-scheduler" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.842828 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b29e862-8b36-45b1-a23a-534083def903" containerName="nova-scheduler-scheduler" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.843485 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 16:27:58 crc kubenswrapper[4998]: W1203 16:27:58.846934 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff8e487a_c00a_47a8_997f_6a0376aa873d.slice/crio-2a9568279cea3e4bea4747d70acb26737c5862ac66c591df673d244c5f95f13a WatchSource:0}: Error finding container 2a9568279cea3e4bea4747d70acb26737c5862ac66c591df673d244c5f95f13a: Status 404 returned error can't find the container with id 2a9568279cea3e4bea4747d70acb26737c5862ac66c591df673d244c5f95f13a Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.847109 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.849842 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.860628 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.969835 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fjlk\" (UniqueName: \"kubernetes.io/projected/63806609-762c-4623-b4e3-55b90b912b5f-kube-api-access-9fjlk\") pod \"nova-scheduler-0\" (UID: \"63806609-762c-4623-b4e3-55b90b912b5f\") " pod="openstack/nova-scheduler-0" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.970000 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63806609-762c-4623-b4e3-55b90b912b5f-config-data\") pod \"nova-scheduler-0\" (UID: \"63806609-762c-4623-b4e3-55b90b912b5f\") " pod="openstack/nova-scheduler-0" Dec 03 16:27:58 crc kubenswrapper[4998]: I1203 16:27:58.970048 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63806609-762c-4623-b4e3-55b90b912b5f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"63806609-762c-4623-b4e3-55b90b912b5f\") " pod="openstack/nova-scheduler-0" Dec 03 16:27:59 crc kubenswrapper[4998]: I1203 16:27:59.071606 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63806609-762c-4623-b4e3-55b90b912b5f-config-data\") pod \"nova-scheduler-0\" (UID: \"63806609-762c-4623-b4e3-55b90b912b5f\") " pod="openstack/nova-scheduler-0" Dec 03 16:27:59 crc kubenswrapper[4998]: I1203 16:27:59.071969 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63806609-762c-4623-b4e3-55b90b912b5f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"63806609-762c-4623-b4e3-55b90b912b5f\") " pod="openstack/nova-scheduler-0" Dec 03 16:27:59 crc kubenswrapper[4998]: I1203 16:27:59.072099 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fjlk\" (UniqueName: \"kubernetes.io/projected/63806609-762c-4623-b4e3-55b90b912b5f-kube-api-access-9fjlk\") pod \"nova-scheduler-0\" (UID: \"63806609-762c-4623-b4e3-55b90b912b5f\") " pod="openstack/nova-scheduler-0" Dec 03 16:27:59 crc kubenswrapper[4998]: I1203 16:27:59.075390 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63806609-762c-4623-b4e3-55b90b912b5f-config-data\") pod \"nova-scheduler-0\" (UID: \"63806609-762c-4623-b4e3-55b90b912b5f\") " pod="openstack/nova-scheduler-0" Dec 03 16:27:59 crc kubenswrapper[4998]: I1203 16:27:59.075506 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63806609-762c-4623-b4e3-55b90b912b5f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"63806609-762c-4623-b4e3-55b90b912b5f\") " pod="openstack/nova-scheduler-0" Dec 03 16:27:59 crc kubenswrapper[4998]: I1203 16:27:59.091935 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fjlk\" (UniqueName: \"kubernetes.io/projected/63806609-762c-4623-b4e3-55b90b912b5f-kube-api-access-9fjlk\") pod \"nova-scheduler-0\" (UID: \"63806609-762c-4623-b4e3-55b90b912b5f\") " pod="openstack/nova-scheduler-0" Dec 03 16:27:59 crc kubenswrapper[4998]: I1203 16:27:59.169572 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 16:27:59 crc kubenswrapper[4998]: I1203 16:27:59.689179 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b29e862-8b36-45b1-a23a-534083def903" path="/var/lib/kubelet/pods/2b29e862-8b36-45b1-a23a-534083def903/volumes" Dec 03 16:27:59 crc kubenswrapper[4998]: I1203 16:27:59.690023 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd010875-398a-48fc-8fa5-569126c4f67f" path="/var/lib/kubelet/pods/cd010875-398a-48fc-8fa5-569126c4f67f/volumes" Dec 03 16:27:59 crc kubenswrapper[4998]: W1203 16:27:59.697675 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63806609_762c_4623_b4e3_55b90b912b5f.slice/crio-a3522d802dc0cd4b00af4deea6982f7f3f5ee876e4a0b8e64a460e204704718d WatchSource:0}: Error finding container a3522d802dc0cd4b00af4deea6982f7f3f5ee876e4a0b8e64a460e204704718d: Status 404 returned error can't find the container with id a3522d802dc0cd4b00af4deea6982f7f3f5ee876e4a0b8e64a460e204704718d Dec 03 16:27:59 crc kubenswrapper[4998]: I1203 16:27:59.711948 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 16:27:59 crc kubenswrapper[4998]: I1203 16:27:59.775933 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"63806609-762c-4623-b4e3-55b90b912b5f","Type":"ContainerStarted","Data":"a3522d802dc0cd4b00af4deea6982f7f3f5ee876e4a0b8e64a460e204704718d"} Dec 03 16:27:59 crc kubenswrapper[4998]: I1203 16:27:59.778474 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ff8e487a-c00a-47a8-997f-6a0376aa873d","Type":"ContainerStarted","Data":"f50dfa105c3067ec62f6dff6a3f690eb80e5b975c967c5e00456221ef11a476c"} Dec 03 16:27:59 crc kubenswrapper[4998]: I1203 16:27:59.778497 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ff8e487a-c00a-47a8-997f-6a0376aa873d","Type":"ContainerStarted","Data":"e4fb39af0bff0724787173d250691c1d8a8d7bb1cb27e36b233d22419bc6646c"} Dec 03 16:27:59 crc kubenswrapper[4998]: I1203 16:27:59.778507 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ff8e487a-c00a-47a8-997f-6a0376aa873d","Type":"ContainerStarted","Data":"2a9568279cea3e4bea4747d70acb26737c5862ac66c591df673d244c5f95f13a"} Dec 03 16:27:59 crc kubenswrapper[4998]: I1203 16:27:59.807448 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.807273553 podStartE2EDuration="2.807273553s" podCreationTimestamp="2025-12-03 16:27:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:27:59.805081879 +0000 UTC m=+1458.416782102" watchObservedRunningTime="2025-12-03 16:27:59.807273553 +0000 UTC m=+1458.418973776" Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.581002 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.716519 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8849882f-39f2-45f3-86e0-e195e81f003c-combined-ca-bundle\") pod \"8849882f-39f2-45f3-86e0-e195e81f003c\" (UID: \"8849882f-39f2-45f3-86e0-e195e81f003c\") " Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.716623 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8849882f-39f2-45f3-86e0-e195e81f003c-public-tls-certs\") pod \"8849882f-39f2-45f3-86e0-e195e81f003c\" (UID: \"8849882f-39f2-45f3-86e0-e195e81f003c\") " Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.716663 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8849882f-39f2-45f3-86e0-e195e81f003c-logs\") pod \"8849882f-39f2-45f3-86e0-e195e81f003c\" (UID: \"8849882f-39f2-45f3-86e0-e195e81f003c\") " Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.716705 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4jzq\" (UniqueName: \"kubernetes.io/projected/8849882f-39f2-45f3-86e0-e195e81f003c-kube-api-access-z4jzq\") pod \"8849882f-39f2-45f3-86e0-e195e81f003c\" (UID: \"8849882f-39f2-45f3-86e0-e195e81f003c\") " Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.716916 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8849882f-39f2-45f3-86e0-e195e81f003c-config-data\") pod \"8849882f-39f2-45f3-86e0-e195e81f003c\" (UID: \"8849882f-39f2-45f3-86e0-e195e81f003c\") " Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.716963 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8849882f-39f2-45f3-86e0-e195e81f003c-internal-tls-certs\") pod \"8849882f-39f2-45f3-86e0-e195e81f003c\" (UID: \"8849882f-39f2-45f3-86e0-e195e81f003c\") " Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.718685 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8849882f-39f2-45f3-86e0-e195e81f003c-logs" (OuterVolumeSpecName: "logs") pod "8849882f-39f2-45f3-86e0-e195e81f003c" (UID: "8849882f-39f2-45f3-86e0-e195e81f003c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.722193 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8849882f-39f2-45f3-86e0-e195e81f003c-kube-api-access-z4jzq" (OuterVolumeSpecName: "kube-api-access-z4jzq") pod "8849882f-39f2-45f3-86e0-e195e81f003c" (UID: "8849882f-39f2-45f3-86e0-e195e81f003c"). InnerVolumeSpecName "kube-api-access-z4jzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.766339 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8849882f-39f2-45f3-86e0-e195e81f003c-config-data" (OuterVolumeSpecName: "config-data") pod "8849882f-39f2-45f3-86e0-e195e81f003c" (UID: "8849882f-39f2-45f3-86e0-e195e81f003c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.766969 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8849882f-39f2-45f3-86e0-e195e81f003c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8849882f-39f2-45f3-86e0-e195e81f003c" (UID: "8849882f-39f2-45f3-86e0-e195e81f003c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.772002 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8849882f-39f2-45f3-86e0-e195e81f003c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8849882f-39f2-45f3-86e0-e195e81f003c" (UID: "8849882f-39f2-45f3-86e0-e195e81f003c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.790710 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"63806609-762c-4623-b4e3-55b90b912b5f","Type":"ContainerStarted","Data":"a3d47287eaaa08a7eca1e7fc4edcfd59436b8ce93a4fff5e620f96da158afb3e"} Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.793501 4998 generic.go:334] "Generic (PLEG): container finished" podID="8849882f-39f2-45f3-86e0-e195e81f003c" containerID="e7ec41fba0cfef6c75ae6dc4be59df4fd8de9d824c1354bd6c58740ec7752f7b" exitCode=0 Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.793617 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8849882f-39f2-45f3-86e0-e195e81f003c","Type":"ContainerDied","Data":"e7ec41fba0cfef6c75ae6dc4be59df4fd8de9d824c1354bd6c58740ec7752f7b"} Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.793654 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8849882f-39f2-45f3-86e0-e195e81f003c","Type":"ContainerDied","Data":"28328a7dbb6d8ab1374b82ae7c43bbe30cbcaf371e347f2bba9be0cbbd095253"} Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.793674 4998 scope.go:117] "RemoveContainer" containerID="e7ec41fba0cfef6c75ae6dc4be59df4fd8de9d824c1354bd6c58740ec7752f7b" Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.793885 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.812619 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.8126000270000002 podStartE2EDuration="2.812600027s" podCreationTimestamp="2025-12-03 16:27:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:28:00.80707433 +0000 UTC m=+1459.418774553" watchObservedRunningTime="2025-12-03 16:28:00.812600027 +0000 UTC m=+1459.424300250" Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.819299 4998 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8849882f-39f2-45f3-86e0-e195e81f003c-logs\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.819337 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4jzq\" (UniqueName: \"kubernetes.io/projected/8849882f-39f2-45f3-86e0-e195e81f003c-kube-api-access-z4jzq\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.819349 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8849882f-39f2-45f3-86e0-e195e81f003c-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.819359 4998 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8849882f-39f2-45f3-86e0-e195e81f003c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.819370 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8849882f-39f2-45f3-86e0-e195e81f003c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.819815 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8849882f-39f2-45f3-86e0-e195e81f003c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8849882f-39f2-45f3-86e0-e195e81f003c" (UID: "8849882f-39f2-45f3-86e0-e195e81f003c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.857492 4998 scope.go:117] "RemoveContainer" containerID="5239376b905ac4e009c765940d81a02c25f161d572cbcd0be748e0ae5d6fd989" Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.884149 4998 scope.go:117] "RemoveContainer" containerID="e7ec41fba0cfef6c75ae6dc4be59df4fd8de9d824c1354bd6c58740ec7752f7b" Dec 03 16:28:00 crc kubenswrapper[4998]: E1203 16:28:00.884499 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7ec41fba0cfef6c75ae6dc4be59df4fd8de9d824c1354bd6c58740ec7752f7b\": container with ID starting with e7ec41fba0cfef6c75ae6dc4be59df4fd8de9d824c1354bd6c58740ec7752f7b not found: ID does not exist" containerID="e7ec41fba0cfef6c75ae6dc4be59df4fd8de9d824c1354bd6c58740ec7752f7b" Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.884529 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7ec41fba0cfef6c75ae6dc4be59df4fd8de9d824c1354bd6c58740ec7752f7b"} err="failed to get container status \"e7ec41fba0cfef6c75ae6dc4be59df4fd8de9d824c1354bd6c58740ec7752f7b\": rpc error: code = NotFound desc = could not find container \"e7ec41fba0cfef6c75ae6dc4be59df4fd8de9d824c1354bd6c58740ec7752f7b\": container with ID starting with e7ec41fba0cfef6c75ae6dc4be59df4fd8de9d824c1354bd6c58740ec7752f7b not found: ID does not exist" Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.884550 4998 scope.go:117] "RemoveContainer" containerID="5239376b905ac4e009c765940d81a02c25f161d572cbcd0be748e0ae5d6fd989" Dec 03 16:28:00 crc kubenswrapper[4998]: E1203 16:28:00.885078 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5239376b905ac4e009c765940d81a02c25f161d572cbcd0be748e0ae5d6fd989\": container with ID starting with 5239376b905ac4e009c765940d81a02c25f161d572cbcd0be748e0ae5d6fd989 not found: ID does not exist" containerID="5239376b905ac4e009c765940d81a02c25f161d572cbcd0be748e0ae5d6fd989" Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.885117 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5239376b905ac4e009c765940d81a02c25f161d572cbcd0be748e0ae5d6fd989"} err="failed to get container status \"5239376b905ac4e009c765940d81a02c25f161d572cbcd0be748e0ae5d6fd989\": rpc error: code = NotFound desc = could not find container \"5239376b905ac4e009c765940d81a02c25f161d572cbcd0be748e0ae5d6fd989\": container with ID starting with 5239376b905ac4e009c765940d81a02c25f161d572cbcd0be748e0ae5d6fd989 not found: ID does not exist" Dec 03 16:28:00 crc kubenswrapper[4998]: I1203 16:28:00.920912 4998 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8849882f-39f2-45f3-86e0-e195e81f003c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.133662 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.158712 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.176991 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 16:28:01 crc kubenswrapper[4998]: E1203 16:28:01.177463 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8849882f-39f2-45f3-86e0-e195e81f003c" containerName="nova-api-api" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.177481 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="8849882f-39f2-45f3-86e0-e195e81f003c" containerName="nova-api-api" Dec 03 16:28:01 crc kubenswrapper[4998]: E1203 16:28:01.177497 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8849882f-39f2-45f3-86e0-e195e81f003c" containerName="nova-api-log" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.177505 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="8849882f-39f2-45f3-86e0-e195e81f003c" containerName="nova-api-log" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.177678 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="8849882f-39f2-45f3-86e0-e195e81f003c" containerName="nova-api-log" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.177700 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="8849882f-39f2-45f3-86e0-e195e81f003c" containerName="nova-api-api" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.178850 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.181683 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.182614 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.182995 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.193799 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.230723 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6468949f-85bc-4274-ada0-b2ef4adb0562-config-data\") pod \"nova-api-0\" (UID: \"6468949f-85bc-4274-ada0-b2ef4adb0562\") " pod="openstack/nova-api-0" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.230788 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6468949f-85bc-4274-ada0-b2ef4adb0562-logs\") pod \"nova-api-0\" (UID: \"6468949f-85bc-4274-ada0-b2ef4adb0562\") " pod="openstack/nova-api-0" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.230821 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6468949f-85bc-4274-ada0-b2ef4adb0562-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6468949f-85bc-4274-ada0-b2ef4adb0562\") " pod="openstack/nova-api-0" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.230847 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6468949f-85bc-4274-ada0-b2ef4adb0562-public-tls-certs\") pod \"nova-api-0\" (UID: \"6468949f-85bc-4274-ada0-b2ef4adb0562\") " pod="openstack/nova-api-0" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.230868 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6468949f-85bc-4274-ada0-b2ef4adb0562-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6468949f-85bc-4274-ada0-b2ef4adb0562\") " pod="openstack/nova-api-0" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.232117 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lx9zr\" (UniqueName: \"kubernetes.io/projected/6468949f-85bc-4274-ada0-b2ef4adb0562-kube-api-access-lx9zr\") pod \"nova-api-0\" (UID: \"6468949f-85bc-4274-ada0-b2ef4adb0562\") " pod="openstack/nova-api-0" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.333746 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6468949f-85bc-4274-ada0-b2ef4adb0562-config-data\") pod \"nova-api-0\" (UID: \"6468949f-85bc-4274-ada0-b2ef4adb0562\") " pod="openstack/nova-api-0" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.333857 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6468949f-85bc-4274-ada0-b2ef4adb0562-logs\") pod \"nova-api-0\" (UID: \"6468949f-85bc-4274-ada0-b2ef4adb0562\") " pod="openstack/nova-api-0" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.333896 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6468949f-85bc-4274-ada0-b2ef4adb0562-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6468949f-85bc-4274-ada0-b2ef4adb0562\") " pod="openstack/nova-api-0" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.333926 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6468949f-85bc-4274-ada0-b2ef4adb0562-public-tls-certs\") pod \"nova-api-0\" (UID: \"6468949f-85bc-4274-ada0-b2ef4adb0562\") " pod="openstack/nova-api-0" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.333951 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6468949f-85bc-4274-ada0-b2ef4adb0562-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6468949f-85bc-4274-ada0-b2ef4adb0562\") " pod="openstack/nova-api-0" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.333983 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lx9zr\" (UniqueName: \"kubernetes.io/projected/6468949f-85bc-4274-ada0-b2ef4adb0562-kube-api-access-lx9zr\") pod \"nova-api-0\" (UID: \"6468949f-85bc-4274-ada0-b2ef4adb0562\") " pod="openstack/nova-api-0" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.334922 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6468949f-85bc-4274-ada0-b2ef4adb0562-logs\") pod \"nova-api-0\" (UID: \"6468949f-85bc-4274-ada0-b2ef4adb0562\") " pod="openstack/nova-api-0" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.338280 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6468949f-85bc-4274-ada0-b2ef4adb0562-public-tls-certs\") pod \"nova-api-0\" (UID: \"6468949f-85bc-4274-ada0-b2ef4adb0562\") " pod="openstack/nova-api-0" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.338649 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6468949f-85bc-4274-ada0-b2ef4adb0562-config-data\") pod \"nova-api-0\" (UID: \"6468949f-85bc-4274-ada0-b2ef4adb0562\") " pod="openstack/nova-api-0" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.339704 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6468949f-85bc-4274-ada0-b2ef4adb0562-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6468949f-85bc-4274-ada0-b2ef4adb0562\") " pod="openstack/nova-api-0" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.352417 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6468949f-85bc-4274-ada0-b2ef4adb0562-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6468949f-85bc-4274-ada0-b2ef4adb0562\") " pod="openstack/nova-api-0" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.359334 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lx9zr\" (UniqueName: \"kubernetes.io/projected/6468949f-85bc-4274-ada0-b2ef4adb0562-kube-api-access-lx9zr\") pod \"nova-api-0\" (UID: \"6468949f-85bc-4274-ada0-b2ef4adb0562\") " pod="openstack/nova-api-0" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.566258 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 16:28:01 crc kubenswrapper[4998]: I1203 16:28:01.692315 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8849882f-39f2-45f3-86e0-e195e81f003c" path="/var/lib/kubelet/pods/8849882f-39f2-45f3-86e0-e195e81f003c/volumes" Dec 03 16:28:02 crc kubenswrapper[4998]: I1203 16:28:02.117286 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 16:28:02 crc kubenswrapper[4998]: I1203 16:28:02.824643 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6468949f-85bc-4274-ada0-b2ef4adb0562","Type":"ContainerStarted","Data":"8d70524a97613d9b8e1a05867935e82bdcac6e4316870f4c28a7a75c29c0d661"} Dec 03 16:28:02 crc kubenswrapper[4998]: I1203 16:28:02.824989 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6468949f-85bc-4274-ada0-b2ef4adb0562","Type":"ContainerStarted","Data":"a46e45855f60fe52689809e16083d49f8dac1deefa17260ee61e7d03eacf0c9b"} Dec 03 16:28:03 crc kubenswrapper[4998]: I1203 16:28:03.337133 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 16:28:03 crc kubenswrapper[4998]: I1203 16:28:03.337205 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 16:28:03 crc kubenswrapper[4998]: I1203 16:28:03.840359 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6468949f-85bc-4274-ada0-b2ef4adb0562","Type":"ContainerStarted","Data":"cf6192fb14bf82b9aeec284f5b97aebfe27f825a565d4945bf5df904ae6771cd"} Dec 03 16:28:03 crc kubenswrapper[4998]: I1203 16:28:03.869965 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.86994236 podStartE2EDuration="2.86994236s" podCreationTimestamp="2025-12-03 16:28:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:28:03.863210703 +0000 UTC m=+1462.474910976" watchObservedRunningTime="2025-12-03 16:28:03.86994236 +0000 UTC m=+1462.481642613" Dec 03 16:28:04 crc kubenswrapper[4998]: I1203 16:28:04.170188 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 16:28:08 crc kubenswrapper[4998]: I1203 16:28:08.336478 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 16:28:08 crc kubenswrapper[4998]: I1203 16:28:08.337059 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 16:28:09 crc kubenswrapper[4998]: I1203 16:28:09.170726 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 16:28:09 crc kubenswrapper[4998]: I1203 16:28:09.199489 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 16:28:09 crc kubenswrapper[4998]: I1203 16:28:09.352913 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ff8e487a-c00a-47a8-997f-6a0376aa873d" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.225:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 16:28:09 crc kubenswrapper[4998]: I1203 16:28:09.352962 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ff8e487a-c00a-47a8-997f-6a0376aa873d" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.225:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 16:28:09 crc kubenswrapper[4998]: I1203 16:28:09.984607 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 16:28:11 crc kubenswrapper[4998]: I1203 16:28:11.566838 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 16:28:11 crc kubenswrapper[4998]: I1203 16:28:11.569062 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 16:28:12 crc kubenswrapper[4998]: I1203 16:28:12.581897 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6468949f-85bc-4274-ada0-b2ef4adb0562" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.227:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 16:28:12 crc kubenswrapper[4998]: I1203 16:28:12.581995 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6468949f-85bc-4274-ada0-b2ef4adb0562" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.227:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 16:28:15 crc kubenswrapper[4998]: I1203 16:28:15.081961 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 16:28:18 crc kubenswrapper[4998]: I1203 16:28:18.341463 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 16:28:18 crc kubenswrapper[4998]: I1203 16:28:18.344684 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 16:28:18 crc kubenswrapper[4998]: I1203 16:28:18.346483 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 16:28:19 crc kubenswrapper[4998]: I1203 16:28:19.029019 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 16:28:21 crc kubenswrapper[4998]: I1203 16:28:21.578992 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 16:28:21 crc kubenswrapper[4998]: I1203 16:28:21.579606 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 16:28:21 crc kubenswrapper[4998]: I1203 16:28:21.583362 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 16:28:21 crc kubenswrapper[4998]: I1203 16:28:21.589274 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 16:28:22 crc kubenswrapper[4998]: I1203 16:28:22.058066 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 16:28:22 crc kubenswrapper[4998]: I1203 16:28:22.073162 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 16:28:27 crc kubenswrapper[4998]: I1203 16:28:27.111737 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:28:27 crc kubenswrapper[4998]: I1203 16:28:27.112137 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:28:27 crc kubenswrapper[4998]: I1203 16:28:27.112204 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:28:27 crc kubenswrapper[4998]: I1203 16:28:27.115104 4998 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379"} pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 16:28:27 crc kubenswrapper[4998]: I1203 16:28:27.115213 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" containerID="cri-o://d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" gracePeriod=600 Dec 03 16:28:27 crc kubenswrapper[4998]: E1203 16:28:27.235746 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:28:28 crc kubenswrapper[4998]: I1203 16:28:28.134556 4998 generic.go:334] "Generic (PLEG): container finished" podID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" exitCode=0 Dec 03 16:28:28 crc kubenswrapper[4998]: I1203 16:28:28.134673 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerDied","Data":"d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379"} Dec 03 16:28:28 crc kubenswrapper[4998]: I1203 16:28:28.134936 4998 scope.go:117] "RemoveContainer" containerID="dc47f739351bd606edc721ca357f0c89b466b48ca6131ea4bbb968b4e4cc02c4" Dec 03 16:28:28 crc kubenswrapper[4998]: I1203 16:28:28.135822 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:28:28 crc kubenswrapper[4998]: E1203 16:28:28.136272 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:28:30 crc kubenswrapper[4998]: I1203 16:28:30.092981 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lmvww"] Dec 03 16:28:30 crc kubenswrapper[4998]: I1203 16:28:30.106098 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lmvww" Dec 03 16:28:30 crc kubenswrapper[4998]: I1203 16:28:30.108992 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lmvww"] Dec 03 16:28:30 crc kubenswrapper[4998]: I1203 16:28:30.200706 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggfpf\" (UniqueName: \"kubernetes.io/projected/bb5f980f-8dd9-4187-9344-e2a8815532af-kube-api-access-ggfpf\") pod \"certified-operators-lmvww\" (UID: \"bb5f980f-8dd9-4187-9344-e2a8815532af\") " pod="openshift-marketplace/certified-operators-lmvww" Dec 03 16:28:30 crc kubenswrapper[4998]: I1203 16:28:30.201121 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb5f980f-8dd9-4187-9344-e2a8815532af-catalog-content\") pod \"certified-operators-lmvww\" (UID: \"bb5f980f-8dd9-4187-9344-e2a8815532af\") " pod="openshift-marketplace/certified-operators-lmvww" Dec 03 16:28:30 crc kubenswrapper[4998]: I1203 16:28:30.201286 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb5f980f-8dd9-4187-9344-e2a8815532af-utilities\") pod \"certified-operators-lmvww\" (UID: \"bb5f980f-8dd9-4187-9344-e2a8815532af\") " pod="openshift-marketplace/certified-operators-lmvww" Dec 03 16:28:30 crc kubenswrapper[4998]: I1203 16:28:30.302167 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggfpf\" (UniqueName: \"kubernetes.io/projected/bb5f980f-8dd9-4187-9344-e2a8815532af-kube-api-access-ggfpf\") pod \"certified-operators-lmvww\" (UID: \"bb5f980f-8dd9-4187-9344-e2a8815532af\") " pod="openshift-marketplace/certified-operators-lmvww" Dec 03 16:28:30 crc kubenswrapper[4998]: I1203 16:28:30.302289 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb5f980f-8dd9-4187-9344-e2a8815532af-catalog-content\") pod \"certified-operators-lmvww\" (UID: \"bb5f980f-8dd9-4187-9344-e2a8815532af\") " pod="openshift-marketplace/certified-operators-lmvww" Dec 03 16:28:30 crc kubenswrapper[4998]: I1203 16:28:30.302327 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb5f980f-8dd9-4187-9344-e2a8815532af-utilities\") pod \"certified-operators-lmvww\" (UID: \"bb5f980f-8dd9-4187-9344-e2a8815532af\") " pod="openshift-marketplace/certified-operators-lmvww" Dec 03 16:28:30 crc kubenswrapper[4998]: I1203 16:28:30.302794 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb5f980f-8dd9-4187-9344-e2a8815532af-utilities\") pod \"certified-operators-lmvww\" (UID: \"bb5f980f-8dd9-4187-9344-e2a8815532af\") " pod="openshift-marketplace/certified-operators-lmvww" Dec 03 16:28:30 crc kubenswrapper[4998]: I1203 16:28:30.302935 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb5f980f-8dd9-4187-9344-e2a8815532af-catalog-content\") pod \"certified-operators-lmvww\" (UID: \"bb5f980f-8dd9-4187-9344-e2a8815532af\") " pod="openshift-marketplace/certified-operators-lmvww" Dec 03 16:28:30 crc kubenswrapper[4998]: I1203 16:28:30.325144 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggfpf\" (UniqueName: \"kubernetes.io/projected/bb5f980f-8dd9-4187-9344-e2a8815532af-kube-api-access-ggfpf\") pod \"certified-operators-lmvww\" (UID: \"bb5f980f-8dd9-4187-9344-e2a8815532af\") " pod="openshift-marketplace/certified-operators-lmvww" Dec 03 16:28:30 crc kubenswrapper[4998]: I1203 16:28:30.442625 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lmvww" Dec 03 16:28:30 crc kubenswrapper[4998]: I1203 16:28:30.825765 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 16:28:30 crc kubenswrapper[4998]: I1203 16:28:30.985594 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lmvww"] Dec 03 16:28:31 crc kubenswrapper[4998]: I1203 16:28:31.181064 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lmvww" event={"ID":"bb5f980f-8dd9-4187-9344-e2a8815532af","Type":"ContainerStarted","Data":"0ad3e582946a0511a26a23eafb571b98a432604a25e5f1ab591f9ad26acf5e8a"} Dec 03 16:28:31 crc kubenswrapper[4998]: I1203 16:28:31.181298 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lmvww" event={"ID":"bb5f980f-8dd9-4187-9344-e2a8815532af","Type":"ContainerStarted","Data":"05c7cf5533f59ddb77fc858e562c249fe2036025740529dd09367772869ba293"} Dec 03 16:28:31 crc kubenswrapper[4998]: I1203 16:28:31.820016 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 16:28:32 crc kubenswrapper[4998]: I1203 16:28:32.191434 4998 generic.go:334] "Generic (PLEG): container finished" podID="bb5f980f-8dd9-4187-9344-e2a8815532af" containerID="0ad3e582946a0511a26a23eafb571b98a432604a25e5f1ab591f9ad26acf5e8a" exitCode=0 Dec 03 16:28:32 crc kubenswrapper[4998]: I1203 16:28:32.191692 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lmvww" event={"ID":"bb5f980f-8dd9-4187-9344-e2a8815532af","Type":"ContainerDied","Data":"0ad3e582946a0511a26a23eafb571b98a432604a25e5f1ab591f9ad26acf5e8a"} Dec 03 16:28:33 crc kubenswrapper[4998]: I1203 16:28:33.202681 4998 generic.go:334] "Generic (PLEG): container finished" podID="bb5f980f-8dd9-4187-9344-e2a8815532af" containerID="1418665249c62e21bf8db8a7fecf3ef48c2b2427c13a77e95c9bec027e115096" exitCode=0 Dec 03 16:28:33 crc kubenswrapper[4998]: I1203 16:28:33.202720 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lmvww" event={"ID":"bb5f980f-8dd9-4187-9344-e2a8815532af","Type":"ContainerDied","Data":"1418665249c62e21bf8db8a7fecf3ef48c2b2427c13a77e95c9bec027e115096"} Dec 03 16:28:34 crc kubenswrapper[4998]: I1203 16:28:34.213804 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lmvww" event={"ID":"bb5f980f-8dd9-4187-9344-e2a8815532af","Type":"ContainerStarted","Data":"534ddfbbe0dbb93d634ffc3404eacf5fea2745d1aebc8d7da33e50f540de2f0c"} Dec 03 16:28:34 crc kubenswrapper[4998]: I1203 16:28:34.240714 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lmvww" podStartSLOduration=2.842457902 podStartE2EDuration="4.240696172s" podCreationTimestamp="2025-12-03 16:28:30 +0000 UTC" firstStartedPulling="2025-12-03 16:28:32.193500696 +0000 UTC m=+1490.805200919" lastFinishedPulling="2025-12-03 16:28:33.591738966 +0000 UTC m=+1492.203439189" observedRunningTime="2025-12-03 16:28:34.234927738 +0000 UTC m=+1492.846627961" watchObservedRunningTime="2025-12-03 16:28:34.240696172 +0000 UTC m=+1492.852396395" Dec 03 16:28:34 crc kubenswrapper[4998]: I1203 16:28:34.557663 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="aeb4a971-3c03-457c-b0df-1503701aac5b" containerName="rabbitmq" containerID="cri-o://26aefd457134bd65e1e106463be93313c5b1fefdf9a1b270d124f5ef4d5ce405" gracePeriod=604797 Dec 03 16:28:35 crc kubenswrapper[4998]: I1203 16:28:35.227283 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="e0ef6669-5a8b-462d-9681-925859c86df1" containerName="rabbitmq" containerID="cri-o://1a47cb6169680516ff6125777af3aa66a1003866aa63ca4fe29fe2f67207bf14" gracePeriod=604797 Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.253400 4998 generic.go:334] "Generic (PLEG): container finished" podID="e0ef6669-5a8b-462d-9681-925859c86df1" containerID="1a47cb6169680516ff6125777af3aa66a1003866aa63ca4fe29fe2f67207bf14" exitCode=0 Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.253813 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e0ef6669-5a8b-462d-9681-925859c86df1","Type":"ContainerDied","Data":"1a47cb6169680516ff6125777af3aa66a1003866aa63ca4fe29fe2f67207bf14"} Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.256411 4998 generic.go:334] "Generic (PLEG): container finished" podID="aeb4a971-3c03-457c-b0df-1503701aac5b" containerID="26aefd457134bd65e1e106463be93313c5b1fefdf9a1b270d124f5ef4d5ce405" exitCode=0 Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.256434 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"aeb4a971-3c03-457c-b0df-1503701aac5b","Type":"ContainerDied","Data":"26aefd457134bd65e1e106463be93313c5b1fefdf9a1b270d124f5ef4d5ce405"} Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.256448 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"aeb4a971-3c03-457c-b0df-1503701aac5b","Type":"ContainerDied","Data":"4b6f937511621fe03bbf3a85fbbb4a9d458cac75832cdb13200fa423199deea7"} Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.256457 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b6f937511621fe03bbf3a85fbbb4a9d458cac75832cdb13200fa423199deea7" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.389164 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.397157 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.475236 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdpq5\" (UniqueName: \"kubernetes.io/projected/aeb4a971-3c03-457c-b0df-1503701aac5b-kube-api-access-hdpq5\") pod \"aeb4a971-3c03-457c-b0df-1503701aac5b\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.475282 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aeb4a971-3c03-457c-b0df-1503701aac5b-config-data\") pod \"aeb4a971-3c03-457c-b0df-1503701aac5b\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.475335 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e0ef6669-5a8b-462d-9681-925859c86df1-rabbitmq-confd\") pod \"e0ef6669-5a8b-462d-9681-925859c86df1\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.475357 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/aeb4a971-3c03-457c-b0df-1503701aac5b-plugins-conf\") pod \"aeb4a971-3c03-457c-b0df-1503701aac5b\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.475394 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e0ef6669-5a8b-462d-9681-925859c86df1-plugins-conf\") pod \"e0ef6669-5a8b-462d-9681-925859c86df1\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.475413 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e0ef6669-5a8b-462d-9681-925859c86df1-erlang-cookie-secret\") pod \"e0ef6669-5a8b-462d-9681-925859c86df1\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.475433 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e0ef6669-5a8b-462d-9681-925859c86df1-config-data\") pod \"e0ef6669-5a8b-462d-9681-925859c86df1\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.475452 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/aeb4a971-3c03-457c-b0df-1503701aac5b-rabbitmq-confd\") pod \"aeb4a971-3c03-457c-b0df-1503701aac5b\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.475472 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e0ef6669-5a8b-462d-9681-925859c86df1-rabbitmq-erlang-cookie\") pod \"e0ef6669-5a8b-462d-9681-925859c86df1\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.475495 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/aeb4a971-3c03-457c-b0df-1503701aac5b-pod-info\") pod \"aeb4a971-3c03-457c-b0df-1503701aac5b\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.475525 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e0ef6669-5a8b-462d-9681-925859c86df1-pod-info\") pod \"e0ef6669-5a8b-462d-9681-925859c86df1\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.475544 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/aeb4a971-3c03-457c-b0df-1503701aac5b-rabbitmq-plugins\") pod \"aeb4a971-3c03-457c-b0df-1503701aac5b\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.475587 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bjlg\" (UniqueName: \"kubernetes.io/projected/e0ef6669-5a8b-462d-9681-925859c86df1-kube-api-access-7bjlg\") pod \"e0ef6669-5a8b-462d-9681-925859c86df1\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.475613 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e0ef6669-5a8b-462d-9681-925859c86df1-rabbitmq-plugins\") pod \"e0ef6669-5a8b-462d-9681-925859c86df1\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.475638 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"e0ef6669-5a8b-462d-9681-925859c86df1\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.475656 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e0ef6669-5a8b-462d-9681-925859c86df1-server-conf\") pod \"e0ef6669-5a8b-462d-9681-925859c86df1\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.475702 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/aeb4a971-3c03-457c-b0df-1503701aac5b-server-conf\") pod \"aeb4a971-3c03-457c-b0df-1503701aac5b\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.475731 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/aeb4a971-3c03-457c-b0df-1503701aac5b-rabbitmq-tls\") pod \"aeb4a971-3c03-457c-b0df-1503701aac5b\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.475750 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e0ef6669-5a8b-462d-9681-925859c86df1-rabbitmq-tls\") pod \"e0ef6669-5a8b-462d-9681-925859c86df1\" (UID: \"e0ef6669-5a8b-462d-9681-925859c86df1\") " Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.475785 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/aeb4a971-3c03-457c-b0df-1503701aac5b-rabbitmq-erlang-cookie\") pod \"aeb4a971-3c03-457c-b0df-1503701aac5b\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.475802 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"aeb4a971-3c03-457c-b0df-1503701aac5b\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.475823 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/aeb4a971-3c03-457c-b0df-1503701aac5b-erlang-cookie-secret\") pod \"aeb4a971-3c03-457c-b0df-1503701aac5b\" (UID: \"aeb4a971-3c03-457c-b0df-1503701aac5b\") " Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.477832 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aeb4a971-3c03-457c-b0df-1503701aac5b-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "aeb4a971-3c03-457c-b0df-1503701aac5b" (UID: "aeb4a971-3c03-457c-b0df-1503701aac5b"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.478468 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aeb4a971-3c03-457c-b0df-1503701aac5b-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "aeb4a971-3c03-457c-b0df-1503701aac5b" (UID: "aeb4a971-3c03-457c-b0df-1503701aac5b"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.484544 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aeb4a971-3c03-457c-b0df-1503701aac5b-kube-api-access-hdpq5" (OuterVolumeSpecName: "kube-api-access-hdpq5") pod "aeb4a971-3c03-457c-b0df-1503701aac5b" (UID: "aeb4a971-3c03-457c-b0df-1503701aac5b"). InnerVolumeSpecName "kube-api-access-hdpq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.488183 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0ef6669-5a8b-462d-9681-925859c86df1-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "e0ef6669-5a8b-462d-9681-925859c86df1" (UID: "e0ef6669-5a8b-462d-9681-925859c86df1"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.488498 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0ef6669-5a8b-462d-9681-925859c86df1-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "e0ef6669-5a8b-462d-9681-925859c86df1" (UID: "e0ef6669-5a8b-462d-9681-925859c86df1"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.489518 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aeb4a971-3c03-457c-b0df-1503701aac5b-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "aeb4a971-3c03-457c-b0df-1503701aac5b" (UID: "aeb4a971-3c03-457c-b0df-1503701aac5b"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.489754 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0ef6669-5a8b-462d-9681-925859c86df1-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "e0ef6669-5a8b-462d-9681-925859c86df1" (UID: "e0ef6669-5a8b-462d-9681-925859c86df1"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.487753 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0ef6669-5a8b-462d-9681-925859c86df1-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "e0ef6669-5a8b-462d-9681-925859c86df1" (UID: "e0ef6669-5a8b-462d-9681-925859c86df1"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.491846 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeb4a971-3c03-457c-b0df-1503701aac5b-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "aeb4a971-3c03-457c-b0df-1503701aac5b" (UID: "aeb4a971-3c03-457c-b0df-1503701aac5b"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.501241 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0ef6669-5a8b-462d-9681-925859c86df1-kube-api-access-7bjlg" (OuterVolumeSpecName: "kube-api-access-7bjlg") pod "e0ef6669-5a8b-462d-9681-925859c86df1" (UID: "e0ef6669-5a8b-462d-9681-925859c86df1"). InnerVolumeSpecName "kube-api-access-7bjlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.502232 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/aeb4a971-3c03-457c-b0df-1503701aac5b-pod-info" (OuterVolumeSpecName: "pod-info") pod "aeb4a971-3c03-457c-b0df-1503701aac5b" (UID: "aeb4a971-3c03-457c-b0df-1503701aac5b"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.506901 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/e0ef6669-5a8b-462d-9681-925859c86df1-pod-info" (OuterVolumeSpecName: "pod-info") pod "e0ef6669-5a8b-462d-9681-925859c86df1" (UID: "e0ef6669-5a8b-462d-9681-925859c86df1"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.508178 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "e0ef6669-5a8b-462d-9681-925859c86df1" (UID: "e0ef6669-5a8b-462d-9681-925859c86df1"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.509446 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aeb4a971-3c03-457c-b0df-1503701aac5b-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "aeb4a971-3c03-457c-b0df-1503701aac5b" (UID: "aeb4a971-3c03-457c-b0df-1503701aac5b"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.510193 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0ef6669-5a8b-462d-9681-925859c86df1-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "e0ef6669-5a8b-462d-9681-925859c86df1" (UID: "e0ef6669-5a8b-462d-9681-925859c86df1"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.510379 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "persistence") pod "aeb4a971-3c03-457c-b0df-1503701aac5b" (UID: "aeb4a971-3c03-457c-b0df-1503701aac5b"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.530301 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aeb4a971-3c03-457c-b0df-1503701aac5b-config-data" (OuterVolumeSpecName: "config-data") pod "aeb4a971-3c03-457c-b0df-1503701aac5b" (UID: "aeb4a971-3c03-457c-b0df-1503701aac5b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.573375 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0ef6669-5a8b-462d-9681-925859c86df1-config-data" (OuterVolumeSpecName: "config-data") pod "e0ef6669-5a8b-462d-9681-925859c86df1" (UID: "e0ef6669-5a8b-462d-9681-925859c86df1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.576048 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aeb4a971-3c03-457c-b0df-1503701aac5b-server-conf" (OuterVolumeSpecName: "server-conf") pod "aeb4a971-3c03-457c-b0df-1503701aac5b" (UID: "aeb4a971-3c03-457c-b0df-1503701aac5b"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.578587 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdpq5\" (UniqueName: \"kubernetes.io/projected/aeb4a971-3c03-457c-b0df-1503701aac5b-kube-api-access-hdpq5\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.579585 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aeb4a971-3c03-457c-b0df-1503701aac5b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.579698 4998 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/aeb4a971-3c03-457c-b0df-1503701aac5b-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.579824 4998 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e0ef6669-5a8b-462d-9681-925859c86df1-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.579918 4998 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e0ef6669-5a8b-462d-9681-925859c86df1-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.580003 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e0ef6669-5a8b-462d-9681-925859c86df1-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.580084 4998 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e0ef6669-5a8b-462d-9681-925859c86df1-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.580155 4998 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/aeb4a971-3c03-457c-b0df-1503701aac5b-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.580225 4998 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e0ef6669-5a8b-462d-9681-925859c86df1-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.580298 4998 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/aeb4a971-3c03-457c-b0df-1503701aac5b-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.580368 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bjlg\" (UniqueName: \"kubernetes.io/projected/e0ef6669-5a8b-462d-9681-925859c86df1-kube-api-access-7bjlg\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.580467 4998 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e0ef6669-5a8b-462d-9681-925859c86df1-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.580578 4998 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.580655 4998 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/aeb4a971-3c03-457c-b0df-1503701aac5b-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.580791 4998 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/aeb4a971-3c03-457c-b0df-1503701aac5b-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.580868 4998 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e0ef6669-5a8b-462d-9681-925859c86df1-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.580964 4998 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.581039 4998 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/aeb4a971-3c03-457c-b0df-1503701aac5b-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.581117 4998 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/aeb4a971-3c03-457c-b0df-1503701aac5b-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.596246 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0ef6669-5a8b-462d-9681-925859c86df1-server-conf" (OuterVolumeSpecName: "server-conf") pod "e0ef6669-5a8b-462d-9681-925859c86df1" (UID: "e0ef6669-5a8b-462d-9681-925859c86df1"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.608934 4998 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.619632 4998 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.660357 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aeb4a971-3c03-457c-b0df-1503701aac5b-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "aeb4a971-3c03-457c-b0df-1503701aac5b" (UID: "aeb4a971-3c03-457c-b0df-1503701aac5b"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.661103 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0ef6669-5a8b-462d-9681-925859c86df1-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "e0ef6669-5a8b-462d-9681-925859c86df1" (UID: "e0ef6669-5a8b-462d-9681-925859c86df1"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.683208 4998 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.683245 4998 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e0ef6669-5a8b-462d-9681-925859c86df1-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.683256 4998 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/aeb4a971-3c03-457c-b0df-1503701aac5b-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.683265 4998 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:37 crc kubenswrapper[4998]: I1203 16:28:37.683274 4998 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e0ef6669-5a8b-462d-9681-925859c86df1-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.268012 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e0ef6669-5a8b-462d-9681-925859c86df1","Type":"ContainerDied","Data":"7374ae0bc9a1d5462c5c581d373c9fc36e39a83f16d0e21c5e3d77f0bb83eb1d"} Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.268924 4998 scope.go:117] "RemoveContainer" containerID="1a47cb6169680516ff6125777af3aa66a1003866aa63ca4fe29fe2f67207bf14" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.268116 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.268026 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.308340 4998 scope.go:117] "RemoveContainer" containerID="a3f53d09f4c25f953ec5e8b2e0a067502b460ce0068895898828046c063bc5c9" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.314690 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.359207 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.429369 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.451303 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.463901 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 16:28:38 crc kubenswrapper[4998]: E1203 16:28:38.465204 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0ef6669-5a8b-462d-9681-925859c86df1" containerName="setup-container" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.465303 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0ef6669-5a8b-462d-9681-925859c86df1" containerName="setup-container" Dec 03 16:28:38 crc kubenswrapper[4998]: E1203 16:28:38.465369 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0ef6669-5a8b-462d-9681-925859c86df1" containerName="rabbitmq" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.465464 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0ef6669-5a8b-462d-9681-925859c86df1" containerName="rabbitmq" Dec 03 16:28:38 crc kubenswrapper[4998]: E1203 16:28:38.465545 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb4a971-3c03-457c-b0df-1503701aac5b" containerName="rabbitmq" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.465612 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb4a971-3c03-457c-b0df-1503701aac5b" containerName="rabbitmq" Dec 03 16:28:38 crc kubenswrapper[4998]: E1203 16:28:38.465688 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeb4a971-3c03-457c-b0df-1503701aac5b" containerName="setup-container" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.465747 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeb4a971-3c03-457c-b0df-1503701aac5b" containerName="setup-container" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.466049 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeb4a971-3c03-457c-b0df-1503701aac5b" containerName="rabbitmq" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.466157 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0ef6669-5a8b-462d-9681-925859c86df1" containerName="rabbitmq" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.467441 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.470777 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.470868 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.470937 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.471125 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.471122 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.471135 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.471816 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-hpxvh" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.483241 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.485054 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.487706 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.488454 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-74b2s" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.488669 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.488861 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.490622 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.490784 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.491137 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.514450 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.526105 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.601888 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.601936 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.601988 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.602072 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6fvs\" (UniqueName: \"kubernetes.io/projected/07b66b88-1162-444d-88d7-13dba5be3ce7-kube-api-access-x6fvs\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.602119 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.602155 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.602193 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.602230 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/07b66b88-1162-444d-88d7-13dba5be3ce7-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.602255 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/07b66b88-1162-444d-88d7-13dba5be3ce7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.602296 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.602339 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/07b66b88-1162-444d-88d7-13dba5be3ce7-config-data\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.602366 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.602472 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/07b66b88-1162-444d-88d7-13dba5be3ce7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.602527 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgxgd\" (UniqueName: \"kubernetes.io/projected/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-kube-api-access-lgxgd\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.602590 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/07b66b88-1162-444d-88d7-13dba5be3ce7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.602712 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/07b66b88-1162-444d-88d7-13dba5be3ce7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.602741 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.603301 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.603344 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/07b66b88-1162-444d-88d7-13dba5be3ce7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.603382 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/07b66b88-1162-444d-88d7-13dba5be3ce7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.603490 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/07b66b88-1162-444d-88d7-13dba5be3ce7-server-conf\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.603519 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.705703 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.705788 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6fvs\" (UniqueName: \"kubernetes.io/projected/07b66b88-1162-444d-88d7-13dba5be3ce7-kube-api-access-x6fvs\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.705828 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.705868 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.705902 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.705934 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/07b66b88-1162-444d-88d7-13dba5be3ce7-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.705956 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/07b66b88-1162-444d-88d7-13dba5be3ce7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.705979 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.706021 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/07b66b88-1162-444d-88d7-13dba5be3ce7-config-data\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.706050 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.706077 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/07b66b88-1162-444d-88d7-13dba5be3ce7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.706097 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgxgd\" (UniqueName: \"kubernetes.io/projected/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-kube-api-access-lgxgd\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.706130 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/07b66b88-1162-444d-88d7-13dba5be3ce7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.706177 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/07b66b88-1162-444d-88d7-13dba5be3ce7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.706197 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.706230 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.706255 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/07b66b88-1162-444d-88d7-13dba5be3ce7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.706280 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/07b66b88-1162-444d-88d7-13dba5be3ce7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.706310 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/07b66b88-1162-444d-88d7-13dba5be3ce7-server-conf\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.706328 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.706351 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.706371 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.706510 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.706933 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.707430 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.707718 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/07b66b88-1162-444d-88d7-13dba5be3ce7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.707878 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/07b66b88-1162-444d-88d7-13dba5be3ce7-server-conf\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.708001 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.708208 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/07b66b88-1162-444d-88d7-13dba5be3ce7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.708413 4998 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.708659 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.708691 4998 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.709154 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/07b66b88-1162-444d-88d7-13dba5be3ce7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.709430 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/07b66b88-1162-444d-88d7-13dba5be3ce7-config-data\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.710350 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.711576 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.714438 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.714656 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/07b66b88-1162-444d-88d7-13dba5be3ce7-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.714864 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/07b66b88-1162-444d-88d7-13dba5be3ce7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.718797 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.720322 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/07b66b88-1162-444d-88d7-13dba5be3ce7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.724272 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/07b66b88-1162-444d-88d7-13dba5be3ce7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.733954 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6fvs\" (UniqueName: \"kubernetes.io/projected/07b66b88-1162-444d-88d7-13dba5be3ce7-kube-api-access-x6fvs\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.735226 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgxgd\" (UniqueName: \"kubernetes.io/projected/887e67c1-bf49-4540-b448-ce1f3ae5a5f2-kube-api-access-lgxgd\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.754637 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-server-0\" (UID: \"07b66b88-1162-444d-88d7-13dba5be3ce7\") " pod="openstack/rabbitmq-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.755124 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"887e67c1-bf49-4540-b448-ce1f3ae5a5f2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.791360 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:28:38 crc kubenswrapper[4998]: I1203 16:28:38.809549 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 16:28:39 crc kubenswrapper[4998]: I1203 16:28:39.263744 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 16:28:39 crc kubenswrapper[4998]: I1203 16:28:39.290107 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"887e67c1-bf49-4540-b448-ce1f3ae5a5f2","Type":"ContainerStarted","Data":"e4f9fe10752168240ac4eb72203c7f928090d1dd5b678313dc9f50dab9283660"} Dec 03 16:28:39 crc kubenswrapper[4998]: I1203 16:28:39.361619 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 16:28:39 crc kubenswrapper[4998]: W1203 16:28:39.365749 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07b66b88_1162_444d_88d7_13dba5be3ce7.slice/crio-4c0cd7f68e180b0efa722655edeecf58e844ca96ae84184d161e20dbf8858935 WatchSource:0}: Error finding container 4c0cd7f68e180b0efa722655edeecf58e844ca96ae84184d161e20dbf8858935: Status 404 returned error can't find the container with id 4c0cd7f68e180b0efa722655edeecf58e844ca96ae84184d161e20dbf8858935 Dec 03 16:28:39 crc kubenswrapper[4998]: I1203 16:28:39.677788 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:28:39 crc kubenswrapper[4998]: E1203 16:28:39.678155 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:28:39 crc kubenswrapper[4998]: I1203 16:28:39.696170 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aeb4a971-3c03-457c-b0df-1503701aac5b" path="/var/lib/kubelet/pods/aeb4a971-3c03-457c-b0df-1503701aac5b/volumes" Dec 03 16:28:39 crc kubenswrapper[4998]: I1203 16:28:39.698508 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0ef6669-5a8b-462d-9681-925859c86df1" path="/var/lib/kubelet/pods/e0ef6669-5a8b-462d-9681-925859c86df1/volumes" Dec 03 16:28:40 crc kubenswrapper[4998]: I1203 16:28:40.304369 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"07b66b88-1162-444d-88d7-13dba5be3ce7","Type":"ContainerStarted","Data":"4c0cd7f68e180b0efa722655edeecf58e844ca96ae84184d161e20dbf8858935"} Dec 03 16:28:40 crc kubenswrapper[4998]: I1203 16:28:40.443376 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lmvww" Dec 03 16:28:40 crc kubenswrapper[4998]: I1203 16:28:40.443719 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lmvww" Dec 03 16:28:40 crc kubenswrapper[4998]: I1203 16:28:40.514178 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lmvww" Dec 03 16:28:41 crc kubenswrapper[4998]: I1203 16:28:41.319554 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"07b66b88-1162-444d-88d7-13dba5be3ce7","Type":"ContainerStarted","Data":"512b7699619874ae07767f10668dbc5d4d14e8a7d75fa70934c7d3d1a0251f2d"} Dec 03 16:28:41 crc kubenswrapper[4998]: I1203 16:28:41.389874 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lmvww" Dec 03 16:28:41 crc kubenswrapper[4998]: I1203 16:28:41.447195 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lmvww"] Dec 03 16:28:42 crc kubenswrapper[4998]: I1203 16:28:42.332514 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"887e67c1-bf49-4540-b448-ce1f3ae5a5f2","Type":"ContainerStarted","Data":"eb2fe4bc2b444407f6d72713c5b7462939887831531271e8f4f77982c5f90f89"} Dec 03 16:28:43 crc kubenswrapper[4998]: I1203 16:28:43.345503 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lmvww" podUID="bb5f980f-8dd9-4187-9344-e2a8815532af" containerName="registry-server" containerID="cri-o://534ddfbbe0dbb93d634ffc3404eacf5fea2745d1aebc8d7da33e50f540de2f0c" gracePeriod=2 Dec 03 16:28:43 crc kubenswrapper[4998]: I1203 16:28:43.828967 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lmvww" Dec 03 16:28:43 crc kubenswrapper[4998]: I1203 16:28:43.971035 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggfpf\" (UniqueName: \"kubernetes.io/projected/bb5f980f-8dd9-4187-9344-e2a8815532af-kube-api-access-ggfpf\") pod \"bb5f980f-8dd9-4187-9344-e2a8815532af\" (UID: \"bb5f980f-8dd9-4187-9344-e2a8815532af\") " Dec 03 16:28:43 crc kubenswrapper[4998]: I1203 16:28:43.971122 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb5f980f-8dd9-4187-9344-e2a8815532af-utilities\") pod \"bb5f980f-8dd9-4187-9344-e2a8815532af\" (UID: \"bb5f980f-8dd9-4187-9344-e2a8815532af\") " Dec 03 16:28:43 crc kubenswrapper[4998]: I1203 16:28:43.971583 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb5f980f-8dd9-4187-9344-e2a8815532af-catalog-content\") pod \"bb5f980f-8dd9-4187-9344-e2a8815532af\" (UID: \"bb5f980f-8dd9-4187-9344-e2a8815532af\") " Dec 03 16:28:43 crc kubenswrapper[4998]: I1203 16:28:43.972069 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb5f980f-8dd9-4187-9344-e2a8815532af-utilities" (OuterVolumeSpecName: "utilities") pod "bb5f980f-8dd9-4187-9344-e2a8815532af" (UID: "bb5f980f-8dd9-4187-9344-e2a8815532af"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:28:43 crc kubenswrapper[4998]: I1203 16:28:43.972483 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb5f980f-8dd9-4187-9344-e2a8815532af-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:43 crc kubenswrapper[4998]: I1203 16:28:43.978021 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb5f980f-8dd9-4187-9344-e2a8815532af-kube-api-access-ggfpf" (OuterVolumeSpecName: "kube-api-access-ggfpf") pod "bb5f980f-8dd9-4187-9344-e2a8815532af" (UID: "bb5f980f-8dd9-4187-9344-e2a8815532af"). InnerVolumeSpecName "kube-api-access-ggfpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:28:44 crc kubenswrapper[4998]: I1203 16:28:44.032040 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb5f980f-8dd9-4187-9344-e2a8815532af-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bb5f980f-8dd9-4187-9344-e2a8815532af" (UID: "bb5f980f-8dd9-4187-9344-e2a8815532af"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:28:44 crc kubenswrapper[4998]: I1203 16:28:44.074381 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb5f980f-8dd9-4187-9344-e2a8815532af-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:44 crc kubenswrapper[4998]: I1203 16:28:44.074421 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggfpf\" (UniqueName: \"kubernetes.io/projected/bb5f980f-8dd9-4187-9344-e2a8815532af-kube-api-access-ggfpf\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:44 crc kubenswrapper[4998]: I1203 16:28:44.364249 4998 generic.go:334] "Generic (PLEG): container finished" podID="bb5f980f-8dd9-4187-9344-e2a8815532af" containerID="534ddfbbe0dbb93d634ffc3404eacf5fea2745d1aebc8d7da33e50f540de2f0c" exitCode=0 Dec 03 16:28:44 crc kubenswrapper[4998]: I1203 16:28:44.364457 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lmvww" event={"ID":"bb5f980f-8dd9-4187-9344-e2a8815532af","Type":"ContainerDied","Data":"534ddfbbe0dbb93d634ffc3404eacf5fea2745d1aebc8d7da33e50f540de2f0c"} Dec 03 16:28:44 crc kubenswrapper[4998]: I1203 16:28:44.364635 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lmvww" event={"ID":"bb5f980f-8dd9-4187-9344-e2a8815532af","Type":"ContainerDied","Data":"05c7cf5533f59ddb77fc858e562c249fe2036025740529dd09367772869ba293"} Dec 03 16:28:44 crc kubenswrapper[4998]: I1203 16:28:44.364664 4998 scope.go:117] "RemoveContainer" containerID="534ddfbbe0dbb93d634ffc3404eacf5fea2745d1aebc8d7da33e50f540de2f0c" Dec 03 16:28:44 crc kubenswrapper[4998]: I1203 16:28:44.364732 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lmvww" Dec 03 16:28:44 crc kubenswrapper[4998]: I1203 16:28:44.401392 4998 scope.go:117] "RemoveContainer" containerID="1418665249c62e21bf8db8a7fecf3ef48c2b2427c13a77e95c9bec027e115096" Dec 03 16:28:44 crc kubenswrapper[4998]: I1203 16:28:44.447866 4998 scope.go:117] "RemoveContainer" containerID="0ad3e582946a0511a26a23eafb571b98a432604a25e5f1ab591f9ad26acf5e8a" Dec 03 16:28:44 crc kubenswrapper[4998]: I1203 16:28:44.524959 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lmvww"] Dec 03 16:28:44 crc kubenswrapper[4998]: I1203 16:28:44.530396 4998 scope.go:117] "RemoveContainer" containerID="534ddfbbe0dbb93d634ffc3404eacf5fea2745d1aebc8d7da33e50f540de2f0c" Dec 03 16:28:44 crc kubenswrapper[4998]: E1203 16:28:44.530864 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"534ddfbbe0dbb93d634ffc3404eacf5fea2745d1aebc8d7da33e50f540de2f0c\": container with ID starting with 534ddfbbe0dbb93d634ffc3404eacf5fea2745d1aebc8d7da33e50f540de2f0c not found: ID does not exist" containerID="534ddfbbe0dbb93d634ffc3404eacf5fea2745d1aebc8d7da33e50f540de2f0c" Dec 03 16:28:44 crc kubenswrapper[4998]: I1203 16:28:44.530895 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"534ddfbbe0dbb93d634ffc3404eacf5fea2745d1aebc8d7da33e50f540de2f0c"} err="failed to get container status \"534ddfbbe0dbb93d634ffc3404eacf5fea2745d1aebc8d7da33e50f540de2f0c\": rpc error: code = NotFound desc = could not find container \"534ddfbbe0dbb93d634ffc3404eacf5fea2745d1aebc8d7da33e50f540de2f0c\": container with ID starting with 534ddfbbe0dbb93d634ffc3404eacf5fea2745d1aebc8d7da33e50f540de2f0c not found: ID does not exist" Dec 03 16:28:44 crc kubenswrapper[4998]: I1203 16:28:44.530916 4998 scope.go:117] "RemoveContainer" containerID="1418665249c62e21bf8db8a7fecf3ef48c2b2427c13a77e95c9bec027e115096" Dec 03 16:28:44 crc kubenswrapper[4998]: E1203 16:28:44.531161 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1418665249c62e21bf8db8a7fecf3ef48c2b2427c13a77e95c9bec027e115096\": container with ID starting with 1418665249c62e21bf8db8a7fecf3ef48c2b2427c13a77e95c9bec027e115096 not found: ID does not exist" containerID="1418665249c62e21bf8db8a7fecf3ef48c2b2427c13a77e95c9bec027e115096" Dec 03 16:28:44 crc kubenswrapper[4998]: I1203 16:28:44.531181 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1418665249c62e21bf8db8a7fecf3ef48c2b2427c13a77e95c9bec027e115096"} err="failed to get container status \"1418665249c62e21bf8db8a7fecf3ef48c2b2427c13a77e95c9bec027e115096\": rpc error: code = NotFound desc = could not find container \"1418665249c62e21bf8db8a7fecf3ef48c2b2427c13a77e95c9bec027e115096\": container with ID starting with 1418665249c62e21bf8db8a7fecf3ef48c2b2427c13a77e95c9bec027e115096 not found: ID does not exist" Dec 03 16:28:44 crc kubenswrapper[4998]: I1203 16:28:44.531195 4998 scope.go:117] "RemoveContainer" containerID="0ad3e582946a0511a26a23eafb571b98a432604a25e5f1ab591f9ad26acf5e8a" Dec 03 16:28:44 crc kubenswrapper[4998]: E1203 16:28:44.531497 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ad3e582946a0511a26a23eafb571b98a432604a25e5f1ab591f9ad26acf5e8a\": container with ID starting with 0ad3e582946a0511a26a23eafb571b98a432604a25e5f1ab591f9ad26acf5e8a not found: ID does not exist" containerID="0ad3e582946a0511a26a23eafb571b98a432604a25e5f1ab591f9ad26acf5e8a" Dec 03 16:28:44 crc kubenswrapper[4998]: I1203 16:28:44.531517 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ad3e582946a0511a26a23eafb571b98a432604a25e5f1ab591f9ad26acf5e8a"} err="failed to get container status \"0ad3e582946a0511a26a23eafb571b98a432604a25e5f1ab591f9ad26acf5e8a\": rpc error: code = NotFound desc = could not find container \"0ad3e582946a0511a26a23eafb571b98a432604a25e5f1ab591f9ad26acf5e8a\": container with ID starting with 0ad3e582946a0511a26a23eafb571b98a432604a25e5f1ab591f9ad26acf5e8a not found: ID does not exist" Dec 03 16:28:44 crc kubenswrapper[4998]: I1203 16:28:44.536624 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lmvww"] Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.224729 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5cd67b459-q7s8x"] Dec 03 16:28:45 crc kubenswrapper[4998]: E1203 16:28:45.225160 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb5f980f-8dd9-4187-9344-e2a8815532af" containerName="extract-utilities" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.225175 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb5f980f-8dd9-4187-9344-e2a8815532af" containerName="extract-utilities" Dec 03 16:28:45 crc kubenswrapper[4998]: E1203 16:28:45.225195 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb5f980f-8dd9-4187-9344-e2a8815532af" containerName="registry-server" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.225201 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb5f980f-8dd9-4187-9344-e2a8815532af" containerName="registry-server" Dec 03 16:28:45 crc kubenswrapper[4998]: E1203 16:28:45.225224 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb5f980f-8dd9-4187-9344-e2a8815532af" containerName="extract-content" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.225231 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb5f980f-8dd9-4187-9344-e2a8815532af" containerName="extract-content" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.225402 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb5f980f-8dd9-4187-9344-e2a8815532af" containerName="registry-server" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.226498 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.231343 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.240610 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cd67b459-q7s8x"] Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.311901 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-dns-svc\") pod \"dnsmasq-dns-5cd67b459-q7s8x\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.312271 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-config\") pod \"dnsmasq-dns-5cd67b459-q7s8x\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.312472 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-dns-swift-storage-0\") pod \"dnsmasq-dns-5cd67b459-q7s8x\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.312547 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jn7zx\" (UniqueName: \"kubernetes.io/projected/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-kube-api-access-jn7zx\") pod \"dnsmasq-dns-5cd67b459-q7s8x\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.312609 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-openstack-edpm-ipam\") pod \"dnsmasq-dns-5cd67b459-q7s8x\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.312655 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-ovsdbserver-sb\") pod \"dnsmasq-dns-5cd67b459-q7s8x\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.312688 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-ovsdbserver-nb\") pod \"dnsmasq-dns-5cd67b459-q7s8x\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.414633 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-dns-swift-storage-0\") pod \"dnsmasq-dns-5cd67b459-q7s8x\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.414698 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jn7zx\" (UniqueName: \"kubernetes.io/projected/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-kube-api-access-jn7zx\") pod \"dnsmasq-dns-5cd67b459-q7s8x\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.414777 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-openstack-edpm-ipam\") pod \"dnsmasq-dns-5cd67b459-q7s8x\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.414807 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-ovsdbserver-sb\") pod \"dnsmasq-dns-5cd67b459-q7s8x\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.414849 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-ovsdbserver-nb\") pod \"dnsmasq-dns-5cd67b459-q7s8x\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.414931 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-dns-svc\") pod \"dnsmasq-dns-5cd67b459-q7s8x\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.414966 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-config\") pod \"dnsmasq-dns-5cd67b459-q7s8x\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.415943 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-openstack-edpm-ipam\") pod \"dnsmasq-dns-5cd67b459-q7s8x\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.415992 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-dns-swift-storage-0\") pod \"dnsmasq-dns-5cd67b459-q7s8x\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.416009 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-ovsdbserver-sb\") pod \"dnsmasq-dns-5cd67b459-q7s8x\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.416029 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-dns-svc\") pod \"dnsmasq-dns-5cd67b459-q7s8x\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.416246 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-config\") pod \"dnsmasq-dns-5cd67b459-q7s8x\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.416316 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-ovsdbserver-nb\") pod \"dnsmasq-dns-5cd67b459-q7s8x\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.440122 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jn7zx\" (UniqueName: \"kubernetes.io/projected/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-kube-api-access-jn7zx\") pod \"dnsmasq-dns-5cd67b459-q7s8x\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.551387 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:28:45 crc kubenswrapper[4998]: I1203 16:28:45.692494 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb5f980f-8dd9-4187-9344-e2a8815532af" path="/var/lib/kubelet/pods/bb5f980f-8dd9-4187-9344-e2a8815532af/volumes" Dec 03 16:28:46 crc kubenswrapper[4998]: I1203 16:28:46.037212 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cd67b459-q7s8x"] Dec 03 16:28:46 crc kubenswrapper[4998]: W1203 16:28:46.038325 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee74cf15_834c_4eb5_a4b5_39ac0199c7cb.slice/crio-b1f3af03975bda49f0879d156946f3a22c5abf393d953e9e294cc9f5571feb7c WatchSource:0}: Error finding container b1f3af03975bda49f0879d156946f3a22c5abf393d953e9e294cc9f5571feb7c: Status 404 returned error can't find the container with id b1f3af03975bda49f0879d156946f3a22c5abf393d953e9e294cc9f5571feb7c Dec 03 16:28:46 crc kubenswrapper[4998]: I1203 16:28:46.385489 4998 generic.go:334] "Generic (PLEG): container finished" podID="ee74cf15-834c-4eb5-a4b5-39ac0199c7cb" containerID="1d9500e66faa517a008cee57e465840cf796a2249ef46d6f263ec7473cf94008" exitCode=0 Dec 03 16:28:46 crc kubenswrapper[4998]: I1203 16:28:46.385551 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" event={"ID":"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb","Type":"ContainerDied","Data":"1d9500e66faa517a008cee57e465840cf796a2249ef46d6f263ec7473cf94008"} Dec 03 16:28:46 crc kubenswrapper[4998]: I1203 16:28:46.385789 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" event={"ID":"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb","Type":"ContainerStarted","Data":"b1f3af03975bda49f0879d156946f3a22c5abf393d953e9e294cc9f5571feb7c"} Dec 03 16:28:47 crc kubenswrapper[4998]: I1203 16:28:47.399944 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" event={"ID":"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb","Type":"ContainerStarted","Data":"ef46488ac0322f09448e6321cb72c2f84d863324ee4f40bd7706abf8d6abd88e"} Dec 03 16:28:47 crc kubenswrapper[4998]: I1203 16:28:47.400461 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:28:47 crc kubenswrapper[4998]: I1203 16:28:47.420678 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" podStartSLOduration=2.420657998 podStartE2EDuration="2.420657998s" podCreationTimestamp="2025-12-03 16:28:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:28:47.416679888 +0000 UTC m=+1506.028380111" watchObservedRunningTime="2025-12-03 16:28:47.420657998 +0000 UTC m=+1506.032358221" Dec 03 16:28:51 crc kubenswrapper[4998]: I1203 16:28:51.685957 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:28:51 crc kubenswrapper[4998]: E1203 16:28:51.687918 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:28:55 crc kubenswrapper[4998]: I1203 16:28:55.552824 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:28:55 crc kubenswrapper[4998]: I1203 16:28:55.640593 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcfbd48f-htqmd"] Dec 03 16:28:55 crc kubenswrapper[4998]: I1203 16:28:55.640857 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" podUID="06704c02-6fcd-4257-bc37-981e2f437db2" containerName="dnsmasq-dns" containerID="cri-o://53707713a41b16da049a2c881a7733f518863fb23ce22f106511060a2ba76072" gracePeriod=10 Dec 03 16:28:55 crc kubenswrapper[4998]: I1203 16:28:55.806271 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d7878c9d7-b9xcm"] Dec 03 16:28:55 crc kubenswrapper[4998]: I1203 16:28:55.815412 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" Dec 03 16:28:55 crc kubenswrapper[4998]: I1203 16:28:55.826727 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d7878c9d7-b9xcm"] Dec 03 16:28:55 crc kubenswrapper[4998]: I1203 16:28:55.942035 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b8aca52-64f7-41bc-a1a1-0ae5ef118606-config\") pod \"dnsmasq-dns-d7878c9d7-b9xcm\" (UID: \"3b8aca52-64f7-41bc-a1a1-0ae5ef118606\") " pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" Dec 03 16:28:55 crc kubenswrapper[4998]: I1203 16:28:55.942080 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b8aca52-64f7-41bc-a1a1-0ae5ef118606-ovsdbserver-nb\") pod \"dnsmasq-dns-d7878c9d7-b9xcm\" (UID: \"3b8aca52-64f7-41bc-a1a1-0ae5ef118606\") " pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" Dec 03 16:28:55 crc kubenswrapper[4998]: I1203 16:28:55.942102 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3b8aca52-64f7-41bc-a1a1-0ae5ef118606-dns-swift-storage-0\") pod \"dnsmasq-dns-d7878c9d7-b9xcm\" (UID: \"3b8aca52-64f7-41bc-a1a1-0ae5ef118606\") " pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" Dec 03 16:28:55 crc kubenswrapper[4998]: I1203 16:28:55.942127 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b8aca52-64f7-41bc-a1a1-0ae5ef118606-ovsdbserver-sb\") pod \"dnsmasq-dns-d7878c9d7-b9xcm\" (UID: \"3b8aca52-64f7-41bc-a1a1-0ae5ef118606\") " pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" Dec 03 16:28:55 crc kubenswrapper[4998]: I1203 16:28:55.942153 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b8aca52-64f7-41bc-a1a1-0ae5ef118606-dns-svc\") pod \"dnsmasq-dns-d7878c9d7-b9xcm\" (UID: \"3b8aca52-64f7-41bc-a1a1-0ae5ef118606\") " pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" Dec 03 16:28:55 crc kubenswrapper[4998]: I1203 16:28:55.942254 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3b8aca52-64f7-41bc-a1a1-0ae5ef118606-openstack-edpm-ipam\") pod \"dnsmasq-dns-d7878c9d7-b9xcm\" (UID: \"3b8aca52-64f7-41bc-a1a1-0ae5ef118606\") " pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" Dec 03 16:28:55 crc kubenswrapper[4998]: I1203 16:28:55.942284 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmp8c\" (UniqueName: \"kubernetes.io/projected/3b8aca52-64f7-41bc-a1a1-0ae5ef118606-kube-api-access-tmp8c\") pod \"dnsmasq-dns-d7878c9d7-b9xcm\" (UID: \"3b8aca52-64f7-41bc-a1a1-0ae5ef118606\") " pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.044894 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b8aca52-64f7-41bc-a1a1-0ae5ef118606-ovsdbserver-nb\") pod \"dnsmasq-dns-d7878c9d7-b9xcm\" (UID: \"3b8aca52-64f7-41bc-a1a1-0ae5ef118606\") " pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.044941 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3b8aca52-64f7-41bc-a1a1-0ae5ef118606-dns-swift-storage-0\") pod \"dnsmasq-dns-d7878c9d7-b9xcm\" (UID: \"3b8aca52-64f7-41bc-a1a1-0ae5ef118606\") " pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.044980 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b8aca52-64f7-41bc-a1a1-0ae5ef118606-ovsdbserver-sb\") pod \"dnsmasq-dns-d7878c9d7-b9xcm\" (UID: \"3b8aca52-64f7-41bc-a1a1-0ae5ef118606\") " pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.045015 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b8aca52-64f7-41bc-a1a1-0ae5ef118606-dns-svc\") pod \"dnsmasq-dns-d7878c9d7-b9xcm\" (UID: \"3b8aca52-64f7-41bc-a1a1-0ae5ef118606\") " pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.045155 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3b8aca52-64f7-41bc-a1a1-0ae5ef118606-openstack-edpm-ipam\") pod \"dnsmasq-dns-d7878c9d7-b9xcm\" (UID: \"3b8aca52-64f7-41bc-a1a1-0ae5ef118606\") " pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.045199 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmp8c\" (UniqueName: \"kubernetes.io/projected/3b8aca52-64f7-41bc-a1a1-0ae5ef118606-kube-api-access-tmp8c\") pod \"dnsmasq-dns-d7878c9d7-b9xcm\" (UID: \"3b8aca52-64f7-41bc-a1a1-0ae5ef118606\") " pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.045255 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b8aca52-64f7-41bc-a1a1-0ae5ef118606-config\") pod \"dnsmasq-dns-d7878c9d7-b9xcm\" (UID: \"3b8aca52-64f7-41bc-a1a1-0ae5ef118606\") " pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.046024 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b8aca52-64f7-41bc-a1a1-0ae5ef118606-ovsdbserver-nb\") pod \"dnsmasq-dns-d7878c9d7-b9xcm\" (UID: \"3b8aca52-64f7-41bc-a1a1-0ae5ef118606\") " pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.046053 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3b8aca52-64f7-41bc-a1a1-0ae5ef118606-dns-swift-storage-0\") pod \"dnsmasq-dns-d7878c9d7-b9xcm\" (UID: \"3b8aca52-64f7-41bc-a1a1-0ae5ef118606\") " pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.046123 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b8aca52-64f7-41bc-a1a1-0ae5ef118606-ovsdbserver-sb\") pod \"dnsmasq-dns-d7878c9d7-b9xcm\" (UID: \"3b8aca52-64f7-41bc-a1a1-0ae5ef118606\") " pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.046436 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b8aca52-64f7-41bc-a1a1-0ae5ef118606-dns-svc\") pod \"dnsmasq-dns-d7878c9d7-b9xcm\" (UID: \"3b8aca52-64f7-41bc-a1a1-0ae5ef118606\") " pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.046774 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b8aca52-64f7-41bc-a1a1-0ae5ef118606-config\") pod \"dnsmasq-dns-d7878c9d7-b9xcm\" (UID: \"3b8aca52-64f7-41bc-a1a1-0ae5ef118606\") " pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.047222 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3b8aca52-64f7-41bc-a1a1-0ae5ef118606-openstack-edpm-ipam\") pod \"dnsmasq-dns-d7878c9d7-b9xcm\" (UID: \"3b8aca52-64f7-41bc-a1a1-0ae5ef118606\") " pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.080120 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmp8c\" (UniqueName: \"kubernetes.io/projected/3b8aca52-64f7-41bc-a1a1-0ae5ef118606-kube-api-access-tmp8c\") pod \"dnsmasq-dns-d7878c9d7-b9xcm\" (UID: \"3b8aca52-64f7-41bc-a1a1-0ae5ef118606\") " pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.160727 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.332878 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.455585 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-dns-swift-storage-0\") pod \"06704c02-6fcd-4257-bc37-981e2f437db2\" (UID: \"06704c02-6fcd-4257-bc37-981e2f437db2\") " Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.455841 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-config\") pod \"06704c02-6fcd-4257-bc37-981e2f437db2\" (UID: \"06704c02-6fcd-4257-bc37-981e2f437db2\") " Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.455882 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkldt\" (UniqueName: \"kubernetes.io/projected/06704c02-6fcd-4257-bc37-981e2f437db2-kube-api-access-bkldt\") pod \"06704c02-6fcd-4257-bc37-981e2f437db2\" (UID: \"06704c02-6fcd-4257-bc37-981e2f437db2\") " Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.455914 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-ovsdbserver-sb\") pod \"06704c02-6fcd-4257-bc37-981e2f437db2\" (UID: \"06704c02-6fcd-4257-bc37-981e2f437db2\") " Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.455997 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-ovsdbserver-nb\") pod \"06704c02-6fcd-4257-bc37-981e2f437db2\" (UID: \"06704c02-6fcd-4257-bc37-981e2f437db2\") " Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.456132 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-dns-svc\") pod \"06704c02-6fcd-4257-bc37-981e2f437db2\" (UID: \"06704c02-6fcd-4257-bc37-981e2f437db2\") " Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.467432 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06704c02-6fcd-4257-bc37-981e2f437db2-kube-api-access-bkldt" (OuterVolumeSpecName: "kube-api-access-bkldt") pod "06704c02-6fcd-4257-bc37-981e2f437db2" (UID: "06704c02-6fcd-4257-bc37-981e2f437db2"). InnerVolumeSpecName "kube-api-access-bkldt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.523028 4998 generic.go:334] "Generic (PLEG): container finished" podID="06704c02-6fcd-4257-bc37-981e2f437db2" containerID="53707713a41b16da049a2c881a7733f518863fb23ce22f106511060a2ba76072" exitCode=0 Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.523069 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" event={"ID":"06704c02-6fcd-4257-bc37-981e2f437db2","Type":"ContainerDied","Data":"53707713a41b16da049a2c881a7733f518863fb23ce22f106511060a2ba76072"} Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.523105 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" event={"ID":"06704c02-6fcd-4257-bc37-981e2f437db2","Type":"ContainerDied","Data":"e5eac56ee0ded262548fa9c7566ed2af682f51ef5f0952365a88ec949dd5a993"} Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.523122 4998 scope.go:117] "RemoveContainer" containerID="53707713a41b16da049a2c881a7733f518863fb23ce22f106511060a2ba76072" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.523240 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfbd48f-htqmd" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.523994 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "06704c02-6fcd-4257-bc37-981e2f437db2" (UID: "06704c02-6fcd-4257-bc37-981e2f437db2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.524413 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-config" (OuterVolumeSpecName: "config") pod "06704c02-6fcd-4257-bc37-981e2f437db2" (UID: "06704c02-6fcd-4257-bc37-981e2f437db2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.548398 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "06704c02-6fcd-4257-bc37-981e2f437db2" (UID: "06704c02-6fcd-4257-bc37-981e2f437db2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.558496 4998 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.558535 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkldt\" (UniqueName: \"kubernetes.io/projected/06704c02-6fcd-4257-bc37-981e2f437db2-kube-api-access-bkldt\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.558548 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.558556 4998 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.568347 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "06704c02-6fcd-4257-bc37-981e2f437db2" (UID: "06704c02-6fcd-4257-bc37-981e2f437db2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.569723 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "06704c02-6fcd-4257-bc37-981e2f437db2" (UID: "06704c02-6fcd-4257-bc37-981e2f437db2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.573927 4998 scope.go:117] "RemoveContainer" containerID="7f28edb8b8f2a2d11935910fb6e6c424a273b30cbdec046fe5b1fab87d86d9f6" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.603806 4998 scope.go:117] "RemoveContainer" containerID="53707713a41b16da049a2c881a7733f518863fb23ce22f106511060a2ba76072" Dec 03 16:28:56 crc kubenswrapper[4998]: E1203 16:28:56.604274 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53707713a41b16da049a2c881a7733f518863fb23ce22f106511060a2ba76072\": container with ID starting with 53707713a41b16da049a2c881a7733f518863fb23ce22f106511060a2ba76072 not found: ID does not exist" containerID="53707713a41b16da049a2c881a7733f518863fb23ce22f106511060a2ba76072" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.604311 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53707713a41b16da049a2c881a7733f518863fb23ce22f106511060a2ba76072"} err="failed to get container status \"53707713a41b16da049a2c881a7733f518863fb23ce22f106511060a2ba76072\": rpc error: code = NotFound desc = could not find container \"53707713a41b16da049a2c881a7733f518863fb23ce22f106511060a2ba76072\": container with ID starting with 53707713a41b16da049a2c881a7733f518863fb23ce22f106511060a2ba76072 not found: ID does not exist" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.604337 4998 scope.go:117] "RemoveContainer" containerID="7f28edb8b8f2a2d11935910fb6e6c424a273b30cbdec046fe5b1fab87d86d9f6" Dec 03 16:28:56 crc kubenswrapper[4998]: E1203 16:28:56.606526 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f28edb8b8f2a2d11935910fb6e6c424a273b30cbdec046fe5b1fab87d86d9f6\": container with ID starting with 7f28edb8b8f2a2d11935910fb6e6c424a273b30cbdec046fe5b1fab87d86d9f6 not found: ID does not exist" containerID="7f28edb8b8f2a2d11935910fb6e6c424a273b30cbdec046fe5b1fab87d86d9f6" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.606567 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f28edb8b8f2a2d11935910fb6e6c424a273b30cbdec046fe5b1fab87d86d9f6"} err="failed to get container status \"7f28edb8b8f2a2d11935910fb6e6c424a273b30cbdec046fe5b1fab87d86d9f6\": rpc error: code = NotFound desc = could not find container \"7f28edb8b8f2a2d11935910fb6e6c424a273b30cbdec046fe5b1fab87d86d9f6\": container with ID starting with 7f28edb8b8f2a2d11935910fb6e6c424a273b30cbdec046fe5b1fab87d86d9f6 not found: ID does not exist" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.660827 4998 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.660858 4998 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/06704c02-6fcd-4257-bc37-981e2f437db2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 16:28:56 crc kubenswrapper[4998]: I1203 16:28:56.711723 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d7878c9d7-b9xcm"] Dec 03 16:28:56 crc kubenswrapper[4998]: W1203 16:28:56.714974 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b8aca52_64f7_41bc_a1a1_0ae5ef118606.slice/crio-cdc69e7fc955e7ecc2b12d7d980da157a40c97ce7b92372b4d15fa3d59bfe7f8 WatchSource:0}: Error finding container cdc69e7fc955e7ecc2b12d7d980da157a40c97ce7b92372b4d15fa3d59bfe7f8: Status 404 returned error can't find the container with id cdc69e7fc955e7ecc2b12d7d980da157a40c97ce7b92372b4d15fa3d59bfe7f8 Dec 03 16:28:57 crc kubenswrapper[4998]: I1203 16:28:57.007259 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcfbd48f-htqmd"] Dec 03 16:28:57 crc kubenswrapper[4998]: I1203 16:28:57.020041 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-fcfbd48f-htqmd"] Dec 03 16:28:57 crc kubenswrapper[4998]: I1203 16:28:57.446873 4998 scope.go:117] "RemoveContainer" containerID="3a0b0d4f0a8d7f25f121dff94ba7f15e27e33f0d83a805c785800edd09699873" Dec 03 16:28:57 crc kubenswrapper[4998]: I1203 16:28:57.473118 4998 scope.go:117] "RemoveContainer" containerID="d9927260898b6eed6b91ad5a3c179db33be6ad3449ab3cb52a37f8296becfaed" Dec 03 16:28:57 crc kubenswrapper[4998]: I1203 16:28:57.542416 4998 generic.go:334] "Generic (PLEG): container finished" podID="3b8aca52-64f7-41bc-a1a1-0ae5ef118606" containerID="98a145a2d57fe0693f0b0c0a31d4e154a190e04e4eac54d5605528973179207d" exitCode=0 Dec 03 16:28:57 crc kubenswrapper[4998]: I1203 16:28:57.542492 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" event={"ID":"3b8aca52-64f7-41bc-a1a1-0ae5ef118606","Type":"ContainerDied","Data":"98a145a2d57fe0693f0b0c0a31d4e154a190e04e4eac54d5605528973179207d"} Dec 03 16:28:57 crc kubenswrapper[4998]: I1203 16:28:57.542524 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" event={"ID":"3b8aca52-64f7-41bc-a1a1-0ae5ef118606","Type":"ContainerStarted","Data":"cdc69e7fc955e7ecc2b12d7d980da157a40c97ce7b92372b4d15fa3d59bfe7f8"} Dec 03 16:28:57 crc kubenswrapper[4998]: I1203 16:28:57.693448 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06704c02-6fcd-4257-bc37-981e2f437db2" path="/var/lib/kubelet/pods/06704c02-6fcd-4257-bc37-981e2f437db2/volumes" Dec 03 16:28:58 crc kubenswrapper[4998]: I1203 16:28:58.556611 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" event={"ID":"3b8aca52-64f7-41bc-a1a1-0ae5ef118606","Type":"ContainerStarted","Data":"41950ba94d6424e30c408df0ddd198c2731da71b782a13ed0e8bf89706f4f565"} Dec 03 16:28:58 crc kubenswrapper[4998]: I1203 16:28:58.556977 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" Dec 03 16:28:58 crc kubenswrapper[4998]: I1203 16:28:58.592613 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" podStartSLOduration=3.592586608 podStartE2EDuration="3.592586608s" podCreationTimestamp="2025-12-03 16:28:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:28:58.581641324 +0000 UTC m=+1517.193341577" watchObservedRunningTime="2025-12-03 16:28:58.592586608 +0000 UTC m=+1517.204286861" Dec 03 16:29:03 crc kubenswrapper[4998]: I1203 16:29:03.678795 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:29:03 crc kubenswrapper[4998]: E1203 16:29:03.679606 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.162979 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d7878c9d7-b9xcm" Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.253651 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cd67b459-q7s8x"] Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.255035 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" podUID="ee74cf15-834c-4eb5-a4b5-39ac0199c7cb" containerName="dnsmasq-dns" containerID="cri-o://ef46488ac0322f09448e6321cb72c2f84d863324ee4f40bd7706abf8d6abd88e" gracePeriod=10 Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.647262 4998 generic.go:334] "Generic (PLEG): container finished" podID="ee74cf15-834c-4eb5-a4b5-39ac0199c7cb" containerID="ef46488ac0322f09448e6321cb72c2f84d863324ee4f40bd7706abf8d6abd88e" exitCode=0 Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.647418 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" event={"ID":"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb","Type":"ContainerDied","Data":"ef46488ac0322f09448e6321cb72c2f84d863324ee4f40bd7706abf8d6abd88e"} Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.751174 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.808025 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-openstack-edpm-ipam\") pod \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.808206 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-ovsdbserver-nb\") pod \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.808332 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jn7zx\" (UniqueName: \"kubernetes.io/projected/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-kube-api-access-jn7zx\") pod \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.808358 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-ovsdbserver-sb\") pod \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.808426 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-config\") pod \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.808450 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-dns-svc\") pod \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.808507 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-dns-swift-storage-0\") pod \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\" (UID: \"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb\") " Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.813448 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-kube-api-access-jn7zx" (OuterVolumeSpecName: "kube-api-access-jn7zx") pod "ee74cf15-834c-4eb5-a4b5-39ac0199c7cb" (UID: "ee74cf15-834c-4eb5-a4b5-39ac0199c7cb"). InnerVolumeSpecName "kube-api-access-jn7zx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.860271 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-config" (OuterVolumeSpecName: "config") pod "ee74cf15-834c-4eb5-a4b5-39ac0199c7cb" (UID: "ee74cf15-834c-4eb5-a4b5-39ac0199c7cb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.875678 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ee74cf15-834c-4eb5-a4b5-39ac0199c7cb" (UID: "ee74cf15-834c-4eb5-a4b5-39ac0199c7cb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.875730 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ee74cf15-834c-4eb5-a4b5-39ac0199c7cb" (UID: "ee74cf15-834c-4eb5-a4b5-39ac0199c7cb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.885858 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ee74cf15-834c-4eb5-a4b5-39ac0199c7cb" (UID: "ee74cf15-834c-4eb5-a4b5-39ac0199c7cb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.886113 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ee74cf15-834c-4eb5-a4b5-39ac0199c7cb" (UID: "ee74cf15-834c-4eb5-a4b5-39ac0199c7cb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.890265 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "ee74cf15-834c-4eb5-a4b5-39ac0199c7cb" (UID: "ee74cf15-834c-4eb5-a4b5-39ac0199c7cb"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.914633 4998 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.914670 4998 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.914683 4998 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.914696 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jn7zx\" (UniqueName: \"kubernetes.io/projected/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-kube-api-access-jn7zx\") on node \"crc\" DevicePath \"\"" Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.914709 4998 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.914722 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:29:06 crc kubenswrapper[4998]: I1203 16:29:06.914734 4998 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 16:29:07 crc kubenswrapper[4998]: I1203 16:29:07.662997 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" event={"ID":"ee74cf15-834c-4eb5-a4b5-39ac0199c7cb","Type":"ContainerDied","Data":"b1f3af03975bda49f0879d156946f3a22c5abf393d953e9e294cc9f5571feb7c"} Dec 03 16:29:07 crc kubenswrapper[4998]: I1203 16:29:07.663063 4998 scope.go:117] "RemoveContainer" containerID="ef46488ac0322f09448e6321cb72c2f84d863324ee4f40bd7706abf8d6abd88e" Dec 03 16:29:07 crc kubenswrapper[4998]: I1203 16:29:07.663067 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd67b459-q7s8x" Dec 03 16:29:07 crc kubenswrapper[4998]: I1203 16:29:07.703810 4998 scope.go:117] "RemoveContainer" containerID="1d9500e66faa517a008cee57e465840cf796a2249ef46d6f263ec7473cf94008" Dec 03 16:29:07 crc kubenswrapper[4998]: I1203 16:29:07.708213 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cd67b459-q7s8x"] Dec 03 16:29:07 crc kubenswrapper[4998]: I1203 16:29:07.722992 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5cd67b459-q7s8x"] Dec 03 16:29:09 crc kubenswrapper[4998]: I1203 16:29:09.697663 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee74cf15-834c-4eb5-a4b5-39ac0199c7cb" path="/var/lib/kubelet/pods/ee74cf15-834c-4eb5-a4b5-39ac0199c7cb/volumes" Dec 03 16:29:13 crc kubenswrapper[4998]: I1203 16:29:13.743065 4998 generic.go:334] "Generic (PLEG): container finished" podID="07b66b88-1162-444d-88d7-13dba5be3ce7" containerID="512b7699619874ae07767f10668dbc5d4d14e8a7d75fa70934c7d3d1a0251f2d" exitCode=0 Dec 03 16:29:13 crc kubenswrapper[4998]: I1203 16:29:13.743160 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"07b66b88-1162-444d-88d7-13dba5be3ce7","Type":"ContainerDied","Data":"512b7699619874ae07767f10668dbc5d4d14e8a7d75fa70934c7d3d1a0251f2d"} Dec 03 16:29:14 crc kubenswrapper[4998]: I1203 16:29:14.753155 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"07b66b88-1162-444d-88d7-13dba5be3ce7","Type":"ContainerStarted","Data":"5a9f008de5d189d1680eacfb91573f1f6527577a314ee89f9e9669c431ed76f8"} Dec 03 16:29:14 crc kubenswrapper[4998]: I1203 16:29:14.754568 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 16:29:14 crc kubenswrapper[4998]: I1203 16:29:14.756350 4998 generic.go:334] "Generic (PLEG): container finished" podID="887e67c1-bf49-4540-b448-ce1f3ae5a5f2" containerID="eb2fe4bc2b444407f6d72713c5b7462939887831531271e8f4f77982c5f90f89" exitCode=0 Dec 03 16:29:14 crc kubenswrapper[4998]: I1203 16:29:14.756376 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"887e67c1-bf49-4540-b448-ce1f3ae5a5f2","Type":"ContainerDied","Data":"eb2fe4bc2b444407f6d72713c5b7462939887831531271e8f4f77982c5f90f89"} Dec 03 16:29:14 crc kubenswrapper[4998]: I1203 16:29:14.791549 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.79152876 podStartE2EDuration="36.79152876s" podCreationTimestamp="2025-12-03 16:28:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:29:14.778538795 +0000 UTC m=+1533.390239038" watchObservedRunningTime="2025-12-03 16:29:14.79152876 +0000 UTC m=+1533.403228993" Dec 03 16:29:15 crc kubenswrapper[4998]: I1203 16:29:15.769561 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"887e67c1-bf49-4540-b448-ce1f3ae5a5f2","Type":"ContainerStarted","Data":"902310943d210a2b4c264ee6d04828fc0daef53aa5b78e103b124bcf56a9e855"} Dec 03 16:29:15 crc kubenswrapper[4998]: I1203 16:29:15.771049 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:29:15 crc kubenswrapper[4998]: I1203 16:29:15.800999 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.800973374 podStartE2EDuration="37.800973374s" podCreationTimestamp="2025-12-03 16:28:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:29:15.797394225 +0000 UTC m=+1534.409094478" watchObservedRunningTime="2025-12-03 16:29:15.800973374 +0000 UTC m=+1534.412673627" Dec 03 16:29:16 crc kubenswrapper[4998]: I1203 16:29:16.679134 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:29:16 crc kubenswrapper[4998]: E1203 16:29:16.679385 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:29:17 crc kubenswrapper[4998]: I1203 16:29:17.845282 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ndmc4"] Dec 03 16:29:17 crc kubenswrapper[4998]: E1203 16:29:17.845966 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee74cf15-834c-4eb5-a4b5-39ac0199c7cb" containerName="dnsmasq-dns" Dec 03 16:29:17 crc kubenswrapper[4998]: I1203 16:29:17.845982 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee74cf15-834c-4eb5-a4b5-39ac0199c7cb" containerName="dnsmasq-dns" Dec 03 16:29:17 crc kubenswrapper[4998]: E1203 16:29:17.845996 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee74cf15-834c-4eb5-a4b5-39ac0199c7cb" containerName="init" Dec 03 16:29:17 crc kubenswrapper[4998]: I1203 16:29:17.846003 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee74cf15-834c-4eb5-a4b5-39ac0199c7cb" containerName="init" Dec 03 16:29:17 crc kubenswrapper[4998]: E1203 16:29:17.846045 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06704c02-6fcd-4257-bc37-981e2f437db2" containerName="init" Dec 03 16:29:17 crc kubenswrapper[4998]: I1203 16:29:17.846054 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="06704c02-6fcd-4257-bc37-981e2f437db2" containerName="init" Dec 03 16:29:17 crc kubenswrapper[4998]: E1203 16:29:17.846070 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06704c02-6fcd-4257-bc37-981e2f437db2" containerName="dnsmasq-dns" Dec 03 16:29:17 crc kubenswrapper[4998]: I1203 16:29:17.846078 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="06704c02-6fcd-4257-bc37-981e2f437db2" containerName="dnsmasq-dns" Dec 03 16:29:17 crc kubenswrapper[4998]: I1203 16:29:17.846311 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="06704c02-6fcd-4257-bc37-981e2f437db2" containerName="dnsmasq-dns" Dec 03 16:29:17 crc kubenswrapper[4998]: I1203 16:29:17.846347 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee74cf15-834c-4eb5-a4b5-39ac0199c7cb" containerName="dnsmasq-dns" Dec 03 16:29:17 crc kubenswrapper[4998]: I1203 16:29:17.848156 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ndmc4" Dec 03 16:29:17 crc kubenswrapper[4998]: I1203 16:29:17.868468 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ndmc4"] Dec 03 16:29:17 crc kubenswrapper[4998]: I1203 16:29:17.945482 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sttnt\" (UniqueName: \"kubernetes.io/projected/aa382fce-18c7-440d-9a97-2620bf24965c-kube-api-access-sttnt\") pod \"community-operators-ndmc4\" (UID: \"aa382fce-18c7-440d-9a97-2620bf24965c\") " pod="openshift-marketplace/community-operators-ndmc4" Dec 03 16:29:17 crc kubenswrapper[4998]: I1203 16:29:17.945735 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa382fce-18c7-440d-9a97-2620bf24965c-catalog-content\") pod \"community-operators-ndmc4\" (UID: \"aa382fce-18c7-440d-9a97-2620bf24965c\") " pod="openshift-marketplace/community-operators-ndmc4" Dec 03 16:29:17 crc kubenswrapper[4998]: I1203 16:29:17.945896 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa382fce-18c7-440d-9a97-2620bf24965c-utilities\") pod \"community-operators-ndmc4\" (UID: \"aa382fce-18c7-440d-9a97-2620bf24965c\") " pod="openshift-marketplace/community-operators-ndmc4" Dec 03 16:29:18 crc kubenswrapper[4998]: I1203 16:29:18.048307 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sttnt\" (UniqueName: \"kubernetes.io/projected/aa382fce-18c7-440d-9a97-2620bf24965c-kube-api-access-sttnt\") pod \"community-operators-ndmc4\" (UID: \"aa382fce-18c7-440d-9a97-2620bf24965c\") " pod="openshift-marketplace/community-operators-ndmc4" Dec 03 16:29:18 crc kubenswrapper[4998]: I1203 16:29:18.048541 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa382fce-18c7-440d-9a97-2620bf24965c-catalog-content\") pod \"community-operators-ndmc4\" (UID: \"aa382fce-18c7-440d-9a97-2620bf24965c\") " pod="openshift-marketplace/community-operators-ndmc4" Dec 03 16:29:18 crc kubenswrapper[4998]: I1203 16:29:18.048643 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa382fce-18c7-440d-9a97-2620bf24965c-utilities\") pod \"community-operators-ndmc4\" (UID: \"aa382fce-18c7-440d-9a97-2620bf24965c\") " pod="openshift-marketplace/community-operators-ndmc4" Dec 03 16:29:18 crc kubenswrapper[4998]: I1203 16:29:18.049116 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa382fce-18c7-440d-9a97-2620bf24965c-catalog-content\") pod \"community-operators-ndmc4\" (UID: \"aa382fce-18c7-440d-9a97-2620bf24965c\") " pod="openshift-marketplace/community-operators-ndmc4" Dec 03 16:29:18 crc kubenswrapper[4998]: I1203 16:29:18.049168 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa382fce-18c7-440d-9a97-2620bf24965c-utilities\") pod \"community-operators-ndmc4\" (UID: \"aa382fce-18c7-440d-9a97-2620bf24965c\") " pod="openshift-marketplace/community-operators-ndmc4" Dec 03 16:29:18 crc kubenswrapper[4998]: I1203 16:29:18.067620 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sttnt\" (UniqueName: \"kubernetes.io/projected/aa382fce-18c7-440d-9a97-2620bf24965c-kube-api-access-sttnt\") pod \"community-operators-ndmc4\" (UID: \"aa382fce-18c7-440d-9a97-2620bf24965c\") " pod="openshift-marketplace/community-operators-ndmc4" Dec 03 16:29:18 crc kubenswrapper[4998]: I1203 16:29:18.168633 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ndmc4" Dec 03 16:29:18 crc kubenswrapper[4998]: I1203 16:29:18.658637 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ndmc4"] Dec 03 16:29:18 crc kubenswrapper[4998]: I1203 16:29:18.802591 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ndmc4" event={"ID":"aa382fce-18c7-440d-9a97-2620bf24965c","Type":"ContainerStarted","Data":"9e999ed3219ab44b02859be6b47166f7d63a9d603db70956c1ee7dcbbc995aa5"} Dec 03 16:29:19 crc kubenswrapper[4998]: I1203 16:29:19.813483 4998 generic.go:334] "Generic (PLEG): container finished" podID="aa382fce-18c7-440d-9a97-2620bf24965c" containerID="caf06eb4faaeae38726ae53c8fb8b1a53b73f7c20c2960795695c75409d980c0" exitCode=0 Dec 03 16:29:19 crc kubenswrapper[4998]: I1203 16:29:19.813547 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ndmc4" event={"ID":"aa382fce-18c7-440d-9a97-2620bf24965c","Type":"ContainerDied","Data":"caf06eb4faaeae38726ae53c8fb8b1a53b73f7c20c2960795695c75409d980c0"} Dec 03 16:29:20 crc kubenswrapper[4998]: I1203 16:29:20.824479 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ndmc4" event={"ID":"aa382fce-18c7-440d-9a97-2620bf24965c","Type":"ContainerStarted","Data":"f35882566ece0242901c5bb66d953d33661c7b73c04bd919bdd9b87a7892502b"} Dec 03 16:29:21 crc kubenswrapper[4998]: I1203 16:29:21.836026 4998 generic.go:334] "Generic (PLEG): container finished" podID="aa382fce-18c7-440d-9a97-2620bf24965c" containerID="f35882566ece0242901c5bb66d953d33661c7b73c04bd919bdd9b87a7892502b" exitCode=0 Dec 03 16:29:21 crc kubenswrapper[4998]: I1203 16:29:21.836092 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ndmc4" event={"ID":"aa382fce-18c7-440d-9a97-2620bf24965c","Type":"ContainerDied","Data":"f35882566ece0242901c5bb66d953d33661c7b73c04bd919bdd9b87a7892502b"} Dec 03 16:29:22 crc kubenswrapper[4998]: I1203 16:29:22.851133 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ndmc4" event={"ID":"aa382fce-18c7-440d-9a97-2620bf24965c","Type":"ContainerStarted","Data":"18c9e65069721d23804568e4bb6f6fb3c1c0f9991e70d91354c70aa65e444370"} Dec 03 16:29:22 crc kubenswrapper[4998]: I1203 16:29:22.875981 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ndmc4" podStartSLOduration=3.4674903710000002 podStartE2EDuration="5.875965115s" podCreationTimestamp="2025-12-03 16:29:17 +0000 UTC" firstStartedPulling="2025-12-03 16:29:19.816080612 +0000 UTC m=+1538.427780835" lastFinishedPulling="2025-12-03 16:29:22.224555356 +0000 UTC m=+1540.836255579" observedRunningTime="2025-12-03 16:29:22.867214095 +0000 UTC m=+1541.478914318" watchObservedRunningTime="2025-12-03 16:29:22.875965115 +0000 UTC m=+1541.487665338" Dec 03 16:29:24 crc kubenswrapper[4998]: I1203 16:29:24.620214 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k"] Dec 03 16:29:24 crc kubenswrapper[4998]: I1203 16:29:24.622944 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k" Dec 03 16:29:24 crc kubenswrapper[4998]: I1203 16:29:24.626344 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 16:29:24 crc kubenswrapper[4998]: I1203 16:29:24.627204 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kcj9x" Dec 03 16:29:24 crc kubenswrapper[4998]: I1203 16:29:24.627446 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 16:29:24 crc kubenswrapper[4998]: I1203 16:29:24.627583 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 16:29:24 crc kubenswrapper[4998]: I1203 16:29:24.641136 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k"] Dec 03 16:29:24 crc kubenswrapper[4998]: I1203 16:29:24.779182 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a43a3b83-302e-44ea-9d16-1c0284e8c90f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k\" (UID: \"a43a3b83-302e-44ea-9d16-1c0284e8c90f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k" Dec 03 16:29:24 crc kubenswrapper[4998]: I1203 16:29:24.779335 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a43a3b83-302e-44ea-9d16-1c0284e8c90f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k\" (UID: \"a43a3b83-302e-44ea-9d16-1c0284e8c90f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k" Dec 03 16:29:24 crc kubenswrapper[4998]: I1203 16:29:24.779387 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a43a3b83-302e-44ea-9d16-1c0284e8c90f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k\" (UID: \"a43a3b83-302e-44ea-9d16-1c0284e8c90f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k" Dec 03 16:29:24 crc kubenswrapper[4998]: I1203 16:29:24.780217 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lg84\" (UniqueName: \"kubernetes.io/projected/a43a3b83-302e-44ea-9d16-1c0284e8c90f-kube-api-access-7lg84\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k\" (UID: \"a43a3b83-302e-44ea-9d16-1c0284e8c90f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k" Dec 03 16:29:24 crc kubenswrapper[4998]: I1203 16:29:24.882442 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a43a3b83-302e-44ea-9d16-1c0284e8c90f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k\" (UID: \"a43a3b83-302e-44ea-9d16-1c0284e8c90f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k" Dec 03 16:29:24 crc kubenswrapper[4998]: I1203 16:29:24.882594 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lg84\" (UniqueName: \"kubernetes.io/projected/a43a3b83-302e-44ea-9d16-1c0284e8c90f-kube-api-access-7lg84\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k\" (UID: \"a43a3b83-302e-44ea-9d16-1c0284e8c90f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k" Dec 03 16:29:24 crc kubenswrapper[4998]: I1203 16:29:24.882680 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a43a3b83-302e-44ea-9d16-1c0284e8c90f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k\" (UID: \"a43a3b83-302e-44ea-9d16-1c0284e8c90f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k" Dec 03 16:29:24 crc kubenswrapper[4998]: I1203 16:29:24.882748 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a43a3b83-302e-44ea-9d16-1c0284e8c90f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k\" (UID: \"a43a3b83-302e-44ea-9d16-1c0284e8c90f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k" Dec 03 16:29:24 crc kubenswrapper[4998]: I1203 16:29:24.905879 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a43a3b83-302e-44ea-9d16-1c0284e8c90f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k\" (UID: \"a43a3b83-302e-44ea-9d16-1c0284e8c90f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k" Dec 03 16:29:24 crc kubenswrapper[4998]: I1203 16:29:24.905919 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a43a3b83-302e-44ea-9d16-1c0284e8c90f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k\" (UID: \"a43a3b83-302e-44ea-9d16-1c0284e8c90f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k" Dec 03 16:29:24 crc kubenswrapper[4998]: I1203 16:29:24.906248 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a43a3b83-302e-44ea-9d16-1c0284e8c90f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k\" (UID: \"a43a3b83-302e-44ea-9d16-1c0284e8c90f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k" Dec 03 16:29:24 crc kubenswrapper[4998]: I1203 16:29:24.909174 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lg84\" (UniqueName: \"kubernetes.io/projected/a43a3b83-302e-44ea-9d16-1c0284e8c90f-kube-api-access-7lg84\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k\" (UID: \"a43a3b83-302e-44ea-9d16-1c0284e8c90f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k" Dec 03 16:29:24 crc kubenswrapper[4998]: I1203 16:29:24.949004 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k" Dec 03 16:29:25 crc kubenswrapper[4998]: I1203 16:29:25.609171 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k"] Dec 03 16:29:25 crc kubenswrapper[4998]: W1203 16:29:25.614566 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda43a3b83_302e_44ea_9d16_1c0284e8c90f.slice/crio-7dfdb0145e54607de52cd2463f0256330ab52ab772bf42074790563e4cd3e7e1 WatchSource:0}: Error finding container 7dfdb0145e54607de52cd2463f0256330ab52ab772bf42074790563e4cd3e7e1: Status 404 returned error can't find the container with id 7dfdb0145e54607de52cd2463f0256330ab52ab772bf42074790563e4cd3e7e1 Dec 03 16:29:25 crc kubenswrapper[4998]: I1203 16:29:25.877348 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k" event={"ID":"a43a3b83-302e-44ea-9d16-1c0284e8c90f","Type":"ContainerStarted","Data":"7dfdb0145e54607de52cd2463f0256330ab52ab772bf42074790563e4cd3e7e1"} Dec 03 16:29:28 crc kubenswrapper[4998]: I1203 16:29:28.168949 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ndmc4" Dec 03 16:29:28 crc kubenswrapper[4998]: I1203 16:29:28.169257 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ndmc4" Dec 03 16:29:28 crc kubenswrapper[4998]: I1203 16:29:28.222328 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ndmc4" Dec 03 16:29:28 crc kubenswrapper[4998]: I1203 16:29:28.793949 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 16:29:28 crc kubenswrapper[4998]: I1203 16:29:28.814398 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 16:29:28 crc kubenswrapper[4998]: I1203 16:29:28.998355 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ndmc4" Dec 03 16:29:29 crc kubenswrapper[4998]: I1203 16:29:29.103130 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ndmc4"] Dec 03 16:29:30 crc kubenswrapper[4998]: I1203 16:29:30.678949 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:29:30 crc kubenswrapper[4998]: E1203 16:29:30.680333 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:29:30 crc kubenswrapper[4998]: I1203 16:29:30.962931 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ndmc4" podUID="aa382fce-18c7-440d-9a97-2620bf24965c" containerName="registry-server" containerID="cri-o://18c9e65069721d23804568e4bb6f6fb3c1c0f9991e70d91354c70aa65e444370" gracePeriod=2 Dec 03 16:29:31 crc kubenswrapper[4998]: I1203 16:29:31.973714 4998 generic.go:334] "Generic (PLEG): container finished" podID="aa382fce-18c7-440d-9a97-2620bf24965c" containerID="18c9e65069721d23804568e4bb6f6fb3c1c0f9991e70d91354c70aa65e444370" exitCode=0 Dec 03 16:29:31 crc kubenswrapper[4998]: I1203 16:29:31.973743 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ndmc4" event={"ID":"aa382fce-18c7-440d-9a97-2620bf24965c","Type":"ContainerDied","Data":"18c9e65069721d23804568e4bb6f6fb3c1c0f9991e70d91354c70aa65e444370"} Dec 03 16:29:38 crc kubenswrapper[4998]: E1203 16:29:38.170320 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 18c9e65069721d23804568e4bb6f6fb3c1c0f9991e70d91354c70aa65e444370 is running failed: container process not found" containerID="18c9e65069721d23804568e4bb6f6fb3c1c0f9991e70d91354c70aa65e444370" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 16:29:38 crc kubenswrapper[4998]: E1203 16:29:38.172666 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 18c9e65069721d23804568e4bb6f6fb3c1c0f9991e70d91354c70aa65e444370 is running failed: container process not found" containerID="18c9e65069721d23804568e4bb6f6fb3c1c0f9991e70d91354c70aa65e444370" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 16:29:38 crc kubenswrapper[4998]: E1203 16:29:38.173024 4998 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 18c9e65069721d23804568e4bb6f6fb3c1c0f9991e70d91354c70aa65e444370 is running failed: container process not found" containerID="18c9e65069721d23804568e4bb6f6fb3c1c0f9991e70d91354c70aa65e444370" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 16:29:38 crc kubenswrapper[4998]: E1203 16:29:38.173082 4998 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 18c9e65069721d23804568e4bb6f6fb3c1c0f9991e70d91354c70aa65e444370 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-ndmc4" podUID="aa382fce-18c7-440d-9a97-2620bf24965c" containerName="registry-server" Dec 03 16:29:41 crc kubenswrapper[4998]: I1203 16:29:41.685828 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:29:41 crc kubenswrapper[4998]: E1203 16:29:41.686458 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:29:42 crc kubenswrapper[4998]: E1203 16:29:42.883150 4998 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest" Dec 03 16:29:42 crc kubenswrapper[4998]: E1203 16:29:42.883292 4998 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 03 16:29:42 crc kubenswrapper[4998]: container &Container{Name:repo-setup-edpm-deployment-openstack-edpm-ipam,Image:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,Command:[],Args:[ansible-runner run /runner -p playbook.yaml -i repo-setup-edpm-deployment-openstack-edpm-ipam],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ANSIBLE_VERBOSITY,Value:2,ValueFrom:nil,},EnvVar{Name:RUNNER_PLAYBOOK,Value: Dec 03 16:29:42 crc kubenswrapper[4998]: - hosts: all Dec 03 16:29:42 crc kubenswrapper[4998]: strategy: linear Dec 03 16:29:42 crc kubenswrapper[4998]: tasks: Dec 03 16:29:42 crc kubenswrapper[4998]: - name: Enable podified-repos Dec 03 16:29:42 crc kubenswrapper[4998]: become: true Dec 03 16:29:42 crc kubenswrapper[4998]: ansible.builtin.shell: | Dec 03 16:29:42 crc kubenswrapper[4998]: set -euxo pipefail Dec 03 16:29:42 crc kubenswrapper[4998]: pushd /var/tmp Dec 03 16:29:42 crc kubenswrapper[4998]: curl -sL https://github.com/openstack-k8s-operators/repo-setup/archive/refs/heads/main.tar.gz | tar -xz Dec 03 16:29:42 crc kubenswrapper[4998]: pushd repo-setup-main Dec 03 16:29:42 crc kubenswrapper[4998]: python3 -m venv ./venv Dec 03 16:29:42 crc kubenswrapper[4998]: PBR_VERSION=0.0.0 ./venv/bin/pip install ./ Dec 03 16:29:42 crc kubenswrapper[4998]: ./venv/bin/repo-setup current-podified -b antelope Dec 03 16:29:42 crc kubenswrapper[4998]: popd Dec 03 16:29:42 crc kubenswrapper[4998]: rm -rf repo-setup-main Dec 03 16:29:42 crc kubenswrapper[4998]: Dec 03 16:29:42 crc kubenswrapper[4998]: Dec 03 16:29:42 crc kubenswrapper[4998]: ,ValueFrom:nil,},EnvVar{Name:RUNNER_EXTRA_VARS,Value: Dec 03 16:29:42 crc kubenswrapper[4998]: edpm_override_hosts: openstack-edpm-ipam Dec 03 16:29:42 crc kubenswrapper[4998]: edpm_service_type: repo-setup Dec 03 16:29:42 crc kubenswrapper[4998]: Dec 03 16:29:42 crc kubenswrapper[4998]: Dec 03 16:29:42 crc kubenswrapper[4998]: ,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:repo-setup-combined-ca-bundle,ReadOnly:false,MountPath:/var/lib/openstack/cacerts/repo-setup,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/runner/env/ssh_key,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:inventory,ReadOnly:false,MountPath:/runner/inventory/hosts,SubPath:inventory,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7lg84,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:openstack-aee-default-env,},Optional:*true,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k_openstack(a43a3b83-302e-44ea-9d16-1c0284e8c90f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Dec 03 16:29:42 crc kubenswrapper[4998]: > logger="UnhandledError" Dec 03 16:29:42 crc kubenswrapper[4998]: E1203 16:29:42.884462 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k" podUID="a43a3b83-302e-44ea-9d16-1c0284e8c90f" Dec 03 16:29:42 crc kubenswrapper[4998]: I1203 16:29:42.951914 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ndmc4" Dec 03 16:29:43 crc kubenswrapper[4998]: I1203 16:29:43.095430 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sttnt\" (UniqueName: \"kubernetes.io/projected/aa382fce-18c7-440d-9a97-2620bf24965c-kube-api-access-sttnt\") pod \"aa382fce-18c7-440d-9a97-2620bf24965c\" (UID: \"aa382fce-18c7-440d-9a97-2620bf24965c\") " Dec 03 16:29:43 crc kubenswrapper[4998]: I1203 16:29:43.095622 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa382fce-18c7-440d-9a97-2620bf24965c-catalog-content\") pod \"aa382fce-18c7-440d-9a97-2620bf24965c\" (UID: \"aa382fce-18c7-440d-9a97-2620bf24965c\") " Dec 03 16:29:43 crc kubenswrapper[4998]: I1203 16:29:43.095721 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa382fce-18c7-440d-9a97-2620bf24965c-utilities\") pod \"aa382fce-18c7-440d-9a97-2620bf24965c\" (UID: \"aa382fce-18c7-440d-9a97-2620bf24965c\") " Dec 03 16:29:43 crc kubenswrapper[4998]: I1203 16:29:43.097001 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa382fce-18c7-440d-9a97-2620bf24965c-utilities" (OuterVolumeSpecName: "utilities") pod "aa382fce-18c7-440d-9a97-2620bf24965c" (UID: "aa382fce-18c7-440d-9a97-2620bf24965c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:29:43 crc kubenswrapper[4998]: I1203 16:29:43.104282 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa382fce-18c7-440d-9a97-2620bf24965c-kube-api-access-sttnt" (OuterVolumeSpecName: "kube-api-access-sttnt") pod "aa382fce-18c7-440d-9a97-2620bf24965c" (UID: "aa382fce-18c7-440d-9a97-2620bf24965c"). InnerVolumeSpecName "kube-api-access-sttnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:29:43 crc kubenswrapper[4998]: I1203 16:29:43.112043 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ndmc4" event={"ID":"aa382fce-18c7-440d-9a97-2620bf24965c","Type":"ContainerDied","Data":"9e999ed3219ab44b02859be6b47166f7d63a9d603db70956c1ee7dcbbc995aa5"} Dec 03 16:29:43 crc kubenswrapper[4998]: I1203 16:29:43.112074 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ndmc4" Dec 03 16:29:43 crc kubenswrapper[4998]: I1203 16:29:43.112121 4998 scope.go:117] "RemoveContainer" containerID="18c9e65069721d23804568e4bb6f6fb3c1c0f9991e70d91354c70aa65e444370" Dec 03 16:29:43 crc kubenswrapper[4998]: E1203 16:29:43.114220 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest\\\"\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k" podUID="a43a3b83-302e-44ea-9d16-1c0284e8c90f" Dec 03 16:29:43 crc kubenswrapper[4998]: I1203 16:29:43.138226 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa382fce-18c7-440d-9a97-2620bf24965c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aa382fce-18c7-440d-9a97-2620bf24965c" (UID: "aa382fce-18c7-440d-9a97-2620bf24965c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:29:43 crc kubenswrapper[4998]: I1203 16:29:43.190052 4998 scope.go:117] "RemoveContainer" containerID="f35882566ece0242901c5bb66d953d33661c7b73c04bd919bdd9b87a7892502b" Dec 03 16:29:43 crc kubenswrapper[4998]: I1203 16:29:43.198038 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sttnt\" (UniqueName: \"kubernetes.io/projected/aa382fce-18c7-440d-9a97-2620bf24965c-kube-api-access-sttnt\") on node \"crc\" DevicePath \"\"" Dec 03 16:29:43 crc kubenswrapper[4998]: I1203 16:29:43.198068 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa382fce-18c7-440d-9a97-2620bf24965c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:29:43 crc kubenswrapper[4998]: I1203 16:29:43.198077 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa382fce-18c7-440d-9a97-2620bf24965c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:29:43 crc kubenswrapper[4998]: I1203 16:29:43.219358 4998 scope.go:117] "RemoveContainer" containerID="caf06eb4faaeae38726ae53c8fb8b1a53b73f7c20c2960795695c75409d980c0" Dec 03 16:29:43 crc kubenswrapper[4998]: I1203 16:29:43.453008 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ndmc4"] Dec 03 16:29:43 crc kubenswrapper[4998]: I1203 16:29:43.461856 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ndmc4"] Dec 03 16:29:43 crc kubenswrapper[4998]: I1203 16:29:43.691830 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa382fce-18c7-440d-9a97-2620bf24965c" path="/var/lib/kubelet/pods/aa382fce-18c7-440d-9a97-2620bf24965c/volumes" Dec 03 16:29:54 crc kubenswrapper[4998]: I1203 16:29:54.680781 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:29:54 crc kubenswrapper[4998]: E1203 16:29:54.681840 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:29:55 crc kubenswrapper[4998]: I1203 16:29:55.122913 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 16:29:56 crc kubenswrapper[4998]: I1203 16:29:56.286676 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k" event={"ID":"a43a3b83-302e-44ea-9d16-1c0284e8c90f","Type":"ContainerStarted","Data":"25c378a3fd9fe79c4c31750278708fd6a9a59bfe4478c0a7459195693024c403"} Dec 03 16:29:56 crc kubenswrapper[4998]: I1203 16:29:56.318388 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k" podStartSLOduration=2.814559633 podStartE2EDuration="32.318361353s" podCreationTimestamp="2025-12-03 16:29:24 +0000 UTC" firstStartedPulling="2025-12-03 16:29:25.616924863 +0000 UTC m=+1544.228625096" lastFinishedPulling="2025-12-03 16:29:55.120726593 +0000 UTC m=+1573.732426816" observedRunningTime="2025-12-03 16:29:56.311165563 +0000 UTC m=+1574.922865836" watchObservedRunningTime="2025-12-03 16:29:56.318361353 +0000 UTC m=+1574.930061596" Dec 03 16:29:57 crc kubenswrapper[4998]: I1203 16:29:57.723996 4998 scope.go:117] "RemoveContainer" containerID="26aefd457134bd65e1e106463be93313c5b1fefdf9a1b270d124f5ef4d5ce405" Dec 03 16:29:57 crc kubenswrapper[4998]: I1203 16:29:57.760853 4998 scope.go:117] "RemoveContainer" containerID="ae868c4032d556a63210bdc8ccf9b359af563c9380921b0a12a6d6c67517a6c8" Dec 03 16:29:57 crc kubenswrapper[4998]: I1203 16:29:57.804620 4998 scope.go:117] "RemoveContainer" containerID="b88c2ed698f351d842f6a803ae1df5e84039dc05b78d6abfdee9e43bf12ecd18" Dec 03 16:30:00 crc kubenswrapper[4998]: I1203 16:30:00.166439 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412990-zcxf2"] Dec 03 16:30:00 crc kubenswrapper[4998]: E1203 16:30:00.167223 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa382fce-18c7-440d-9a97-2620bf24965c" containerName="extract-content" Dec 03 16:30:00 crc kubenswrapper[4998]: I1203 16:30:00.167236 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa382fce-18c7-440d-9a97-2620bf24965c" containerName="extract-content" Dec 03 16:30:00 crc kubenswrapper[4998]: E1203 16:30:00.167255 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa382fce-18c7-440d-9a97-2620bf24965c" containerName="registry-server" Dec 03 16:30:00 crc kubenswrapper[4998]: I1203 16:30:00.167260 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa382fce-18c7-440d-9a97-2620bf24965c" containerName="registry-server" Dec 03 16:30:00 crc kubenswrapper[4998]: E1203 16:30:00.167277 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa382fce-18c7-440d-9a97-2620bf24965c" containerName="extract-utilities" Dec 03 16:30:00 crc kubenswrapper[4998]: I1203 16:30:00.167285 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa382fce-18c7-440d-9a97-2620bf24965c" containerName="extract-utilities" Dec 03 16:30:00 crc kubenswrapper[4998]: I1203 16:30:00.167480 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa382fce-18c7-440d-9a97-2620bf24965c" containerName="registry-server" Dec 03 16:30:00 crc kubenswrapper[4998]: I1203 16:30:00.168202 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412990-zcxf2" Dec 03 16:30:00 crc kubenswrapper[4998]: I1203 16:30:00.170621 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 16:30:00 crc kubenswrapper[4998]: I1203 16:30:00.170886 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 16:30:00 crc kubenswrapper[4998]: I1203 16:30:00.178558 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412990-zcxf2"] Dec 03 16:30:00 crc kubenswrapper[4998]: I1203 16:30:00.316327 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjnqt\" (UniqueName: \"kubernetes.io/projected/bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0-kube-api-access-jjnqt\") pod \"collect-profiles-29412990-zcxf2\" (UID: \"bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412990-zcxf2" Dec 03 16:30:00 crc kubenswrapper[4998]: I1203 16:30:00.316617 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0-config-volume\") pod \"collect-profiles-29412990-zcxf2\" (UID: \"bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412990-zcxf2" Dec 03 16:30:00 crc kubenswrapper[4998]: I1203 16:30:00.316658 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0-secret-volume\") pod \"collect-profiles-29412990-zcxf2\" (UID: \"bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412990-zcxf2" Dec 03 16:30:00 crc kubenswrapper[4998]: I1203 16:30:00.419078 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0-config-volume\") pod \"collect-profiles-29412990-zcxf2\" (UID: \"bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412990-zcxf2" Dec 03 16:30:00 crc kubenswrapper[4998]: I1203 16:30:00.419141 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0-secret-volume\") pod \"collect-profiles-29412990-zcxf2\" (UID: \"bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412990-zcxf2" Dec 03 16:30:00 crc kubenswrapper[4998]: I1203 16:30:00.419252 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjnqt\" (UniqueName: \"kubernetes.io/projected/bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0-kube-api-access-jjnqt\") pod \"collect-profiles-29412990-zcxf2\" (UID: \"bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412990-zcxf2" Dec 03 16:30:00 crc kubenswrapper[4998]: I1203 16:30:00.420874 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0-config-volume\") pod \"collect-profiles-29412990-zcxf2\" (UID: \"bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412990-zcxf2" Dec 03 16:30:00 crc kubenswrapper[4998]: I1203 16:30:00.426847 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0-secret-volume\") pod \"collect-profiles-29412990-zcxf2\" (UID: \"bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412990-zcxf2" Dec 03 16:30:00 crc kubenswrapper[4998]: I1203 16:30:00.450749 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjnqt\" (UniqueName: \"kubernetes.io/projected/bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0-kube-api-access-jjnqt\") pod \"collect-profiles-29412990-zcxf2\" (UID: \"bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412990-zcxf2" Dec 03 16:30:00 crc kubenswrapper[4998]: I1203 16:30:00.491493 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412990-zcxf2" Dec 03 16:30:00 crc kubenswrapper[4998]: I1203 16:30:00.941305 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412990-zcxf2"] Dec 03 16:30:00 crc kubenswrapper[4998]: W1203 16:30:00.941893 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd7e24f1_cbc1_4962_8b5b_f1cc1afa2dc0.slice/crio-ed316d2f606531367cac9ff4a30ce61c5ae4f474073e207ab21efc38b1fb9a30 WatchSource:0}: Error finding container ed316d2f606531367cac9ff4a30ce61c5ae4f474073e207ab21efc38b1fb9a30: Status 404 returned error can't find the container with id ed316d2f606531367cac9ff4a30ce61c5ae4f474073e207ab21efc38b1fb9a30 Dec 03 16:30:01 crc kubenswrapper[4998]: I1203 16:30:01.341141 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412990-zcxf2" event={"ID":"bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0","Type":"ContainerStarted","Data":"ed316d2f606531367cac9ff4a30ce61c5ae4f474073e207ab21efc38b1fb9a30"} Dec 03 16:30:02 crc kubenswrapper[4998]: I1203 16:30:02.361543 4998 generic.go:334] "Generic (PLEG): container finished" podID="bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0" containerID="b0f6a6a2769945bfb5e176d55d2f18bbd755e2fc39a356d4b06808ce4c40b542" exitCode=0 Dec 03 16:30:02 crc kubenswrapper[4998]: I1203 16:30:02.361684 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412990-zcxf2" event={"ID":"bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0","Type":"ContainerDied","Data":"b0f6a6a2769945bfb5e176d55d2f18bbd755e2fc39a356d4b06808ce4c40b542"} Dec 03 16:30:03 crc kubenswrapper[4998]: I1203 16:30:03.733884 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412990-zcxf2" Dec 03 16:30:03 crc kubenswrapper[4998]: I1203 16:30:03.889868 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0-secret-volume\") pod \"bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0\" (UID: \"bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0\") " Dec 03 16:30:03 crc kubenswrapper[4998]: I1203 16:30:03.890347 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjnqt\" (UniqueName: \"kubernetes.io/projected/bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0-kube-api-access-jjnqt\") pod \"bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0\" (UID: \"bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0\") " Dec 03 16:30:03 crc kubenswrapper[4998]: I1203 16:30:03.890406 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0-config-volume\") pod \"bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0\" (UID: \"bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0\") " Dec 03 16:30:03 crc kubenswrapper[4998]: I1203 16:30:03.891279 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0-config-volume" (OuterVolumeSpecName: "config-volume") pod "bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0" (UID: "bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:30:03 crc kubenswrapper[4998]: I1203 16:30:03.896806 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0" (UID: "bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:30:03 crc kubenswrapper[4998]: I1203 16:30:03.903893 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0-kube-api-access-jjnqt" (OuterVolumeSpecName: "kube-api-access-jjnqt") pod "bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0" (UID: "bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0"). InnerVolumeSpecName "kube-api-access-jjnqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:30:03 crc kubenswrapper[4998]: I1203 16:30:03.993363 4998 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 16:30:03 crc kubenswrapper[4998]: I1203 16:30:03.993398 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjnqt\" (UniqueName: \"kubernetes.io/projected/bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0-kube-api-access-jjnqt\") on node \"crc\" DevicePath \"\"" Dec 03 16:30:03 crc kubenswrapper[4998]: I1203 16:30:03.993408 4998 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 16:30:04 crc kubenswrapper[4998]: I1203 16:30:04.387594 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412990-zcxf2" event={"ID":"bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0","Type":"ContainerDied","Data":"ed316d2f606531367cac9ff4a30ce61c5ae4f474073e207ab21efc38b1fb9a30"} Dec 03 16:30:04 crc kubenswrapper[4998]: I1203 16:30:04.387644 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed316d2f606531367cac9ff4a30ce61c5ae4f474073e207ab21efc38b1fb9a30" Dec 03 16:30:04 crc kubenswrapper[4998]: I1203 16:30:04.387695 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412990-zcxf2" Dec 03 16:30:08 crc kubenswrapper[4998]: I1203 16:30:08.453019 4998 generic.go:334] "Generic (PLEG): container finished" podID="a43a3b83-302e-44ea-9d16-1c0284e8c90f" containerID="25c378a3fd9fe79c4c31750278708fd6a9a59bfe4478c0a7459195693024c403" exitCode=0 Dec 03 16:30:08 crc kubenswrapper[4998]: I1203 16:30:08.453057 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k" event={"ID":"a43a3b83-302e-44ea-9d16-1c0284e8c90f","Type":"ContainerDied","Data":"25c378a3fd9fe79c4c31750278708fd6a9a59bfe4478c0a7459195693024c403"} Dec 03 16:30:09 crc kubenswrapper[4998]: I1203 16:30:09.677830 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:30:09 crc kubenswrapper[4998]: E1203 16:30:09.678383 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:30:09 crc kubenswrapper[4998]: I1203 16:30:09.946832 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.021916 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lg84\" (UniqueName: \"kubernetes.io/projected/a43a3b83-302e-44ea-9d16-1c0284e8c90f-kube-api-access-7lg84\") pod \"a43a3b83-302e-44ea-9d16-1c0284e8c90f\" (UID: \"a43a3b83-302e-44ea-9d16-1c0284e8c90f\") " Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.022252 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a43a3b83-302e-44ea-9d16-1c0284e8c90f-ssh-key\") pod \"a43a3b83-302e-44ea-9d16-1c0284e8c90f\" (UID: \"a43a3b83-302e-44ea-9d16-1c0284e8c90f\") " Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.022318 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a43a3b83-302e-44ea-9d16-1c0284e8c90f-inventory\") pod \"a43a3b83-302e-44ea-9d16-1c0284e8c90f\" (UID: \"a43a3b83-302e-44ea-9d16-1c0284e8c90f\") " Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.022387 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a43a3b83-302e-44ea-9d16-1c0284e8c90f-repo-setup-combined-ca-bundle\") pod \"a43a3b83-302e-44ea-9d16-1c0284e8c90f\" (UID: \"a43a3b83-302e-44ea-9d16-1c0284e8c90f\") " Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.033716 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a43a3b83-302e-44ea-9d16-1c0284e8c90f-kube-api-access-7lg84" (OuterVolumeSpecName: "kube-api-access-7lg84") pod "a43a3b83-302e-44ea-9d16-1c0284e8c90f" (UID: "a43a3b83-302e-44ea-9d16-1c0284e8c90f"). InnerVolumeSpecName "kube-api-access-7lg84". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.035820 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a43a3b83-302e-44ea-9d16-1c0284e8c90f-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "a43a3b83-302e-44ea-9d16-1c0284e8c90f" (UID: "a43a3b83-302e-44ea-9d16-1c0284e8c90f"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.059046 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a43a3b83-302e-44ea-9d16-1c0284e8c90f-inventory" (OuterVolumeSpecName: "inventory") pod "a43a3b83-302e-44ea-9d16-1c0284e8c90f" (UID: "a43a3b83-302e-44ea-9d16-1c0284e8c90f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.059964 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a43a3b83-302e-44ea-9d16-1c0284e8c90f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a43a3b83-302e-44ea-9d16-1c0284e8c90f" (UID: "a43a3b83-302e-44ea-9d16-1c0284e8c90f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.124959 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lg84\" (UniqueName: \"kubernetes.io/projected/a43a3b83-302e-44ea-9d16-1c0284e8c90f-kube-api-access-7lg84\") on node \"crc\" DevicePath \"\"" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.124986 4998 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a43a3b83-302e-44ea-9d16-1c0284e8c90f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.124995 4998 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a43a3b83-302e-44ea-9d16-1c0284e8c90f-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.125004 4998 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a43a3b83-302e-44ea-9d16-1c0284e8c90f-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.479289 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k" event={"ID":"a43a3b83-302e-44ea-9d16-1c0284e8c90f","Type":"ContainerDied","Data":"7dfdb0145e54607de52cd2463f0256330ab52ab772bf42074790563e4cd3e7e1"} Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.479330 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7dfdb0145e54607de52cd2463f0256330ab52ab772bf42074790563e4cd3e7e1" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.479384 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.562691 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbwwm"] Dec 03 16:30:10 crc kubenswrapper[4998]: E1203 16:30:10.563230 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a43a3b83-302e-44ea-9d16-1c0284e8c90f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.563249 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="a43a3b83-302e-44ea-9d16-1c0284e8c90f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 16:30:10 crc kubenswrapper[4998]: E1203 16:30:10.563281 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0" containerName="collect-profiles" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.563288 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0" containerName="collect-profiles" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.563492 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="a43a3b83-302e-44ea-9d16-1c0284e8c90f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.563517 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0" containerName="collect-profiles" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.564204 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbwwm" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.566533 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kcj9x" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.567805 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.568396 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.568797 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.575037 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbwwm"] Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.634433 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfw59\" (UniqueName: \"kubernetes.io/projected/5311a655-a897-486e-80ce-10a0b8082918-kube-api-access-nfw59\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fbwwm\" (UID: \"5311a655-a897-486e-80ce-10a0b8082918\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbwwm" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.634550 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5311a655-a897-486e-80ce-10a0b8082918-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fbwwm\" (UID: \"5311a655-a897-486e-80ce-10a0b8082918\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbwwm" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.634577 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5311a655-a897-486e-80ce-10a0b8082918-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fbwwm\" (UID: \"5311a655-a897-486e-80ce-10a0b8082918\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbwwm" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.736567 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfw59\" (UniqueName: \"kubernetes.io/projected/5311a655-a897-486e-80ce-10a0b8082918-kube-api-access-nfw59\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fbwwm\" (UID: \"5311a655-a897-486e-80ce-10a0b8082918\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbwwm" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.736709 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5311a655-a897-486e-80ce-10a0b8082918-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fbwwm\" (UID: \"5311a655-a897-486e-80ce-10a0b8082918\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbwwm" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.736748 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5311a655-a897-486e-80ce-10a0b8082918-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fbwwm\" (UID: \"5311a655-a897-486e-80ce-10a0b8082918\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbwwm" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.742179 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5311a655-a897-486e-80ce-10a0b8082918-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fbwwm\" (UID: \"5311a655-a897-486e-80ce-10a0b8082918\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbwwm" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.749303 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5311a655-a897-486e-80ce-10a0b8082918-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fbwwm\" (UID: \"5311a655-a897-486e-80ce-10a0b8082918\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbwwm" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.757083 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfw59\" (UniqueName: \"kubernetes.io/projected/5311a655-a897-486e-80ce-10a0b8082918-kube-api-access-nfw59\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fbwwm\" (UID: \"5311a655-a897-486e-80ce-10a0b8082918\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbwwm" Dec 03 16:30:10 crc kubenswrapper[4998]: I1203 16:30:10.883557 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbwwm" Dec 03 16:30:11 crc kubenswrapper[4998]: I1203 16:30:11.479645 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbwwm"] Dec 03 16:30:11 crc kubenswrapper[4998]: W1203 16:30:11.484925 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5311a655_a897_486e_80ce_10a0b8082918.slice/crio-d89379d42acbfe0c7776fc385ecbae4281a7a441fe32332420631b0a119c9661 WatchSource:0}: Error finding container d89379d42acbfe0c7776fc385ecbae4281a7a441fe32332420631b0a119c9661: Status 404 returned error can't find the container with id d89379d42acbfe0c7776fc385ecbae4281a7a441fe32332420631b0a119c9661 Dec 03 16:30:11 crc kubenswrapper[4998]: I1203 16:30:11.489307 4998 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 16:30:12 crc kubenswrapper[4998]: I1203 16:30:12.508297 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbwwm" event={"ID":"5311a655-a897-486e-80ce-10a0b8082918","Type":"ContainerStarted","Data":"d89379d42acbfe0c7776fc385ecbae4281a7a441fe32332420631b0a119c9661"} Dec 03 16:30:13 crc kubenswrapper[4998]: I1203 16:30:13.522776 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbwwm" event={"ID":"5311a655-a897-486e-80ce-10a0b8082918","Type":"ContainerStarted","Data":"eabea7f9e8e5a606deeffadccfaf9cb4090ab3021198fd4f80b130138f4c167b"} Dec 03 16:30:13 crc kubenswrapper[4998]: I1203 16:30:13.537577 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbwwm" podStartSLOduration=1.803498035 podStartE2EDuration="3.537553884s" podCreationTimestamp="2025-12-03 16:30:10 +0000 UTC" firstStartedPulling="2025-12-03 16:30:11.489000051 +0000 UTC m=+1590.100700284" lastFinishedPulling="2025-12-03 16:30:13.22305591 +0000 UTC m=+1591.834756133" observedRunningTime="2025-12-03 16:30:13.536979539 +0000 UTC m=+1592.148679802" watchObservedRunningTime="2025-12-03 16:30:13.537553884 +0000 UTC m=+1592.149254137" Dec 03 16:30:16 crc kubenswrapper[4998]: I1203 16:30:16.553720 4998 generic.go:334] "Generic (PLEG): container finished" podID="5311a655-a897-486e-80ce-10a0b8082918" containerID="eabea7f9e8e5a606deeffadccfaf9cb4090ab3021198fd4f80b130138f4c167b" exitCode=0 Dec 03 16:30:16 crc kubenswrapper[4998]: I1203 16:30:16.553835 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbwwm" event={"ID":"5311a655-a897-486e-80ce-10a0b8082918","Type":"ContainerDied","Data":"eabea7f9e8e5a606deeffadccfaf9cb4090ab3021198fd4f80b130138f4c167b"} Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.037102 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbwwm" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.107155 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5311a655-a897-486e-80ce-10a0b8082918-inventory\") pod \"5311a655-a897-486e-80ce-10a0b8082918\" (UID: \"5311a655-a897-486e-80ce-10a0b8082918\") " Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.107366 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfw59\" (UniqueName: \"kubernetes.io/projected/5311a655-a897-486e-80ce-10a0b8082918-kube-api-access-nfw59\") pod \"5311a655-a897-486e-80ce-10a0b8082918\" (UID: \"5311a655-a897-486e-80ce-10a0b8082918\") " Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.107539 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5311a655-a897-486e-80ce-10a0b8082918-ssh-key\") pod \"5311a655-a897-486e-80ce-10a0b8082918\" (UID: \"5311a655-a897-486e-80ce-10a0b8082918\") " Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.112801 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5311a655-a897-486e-80ce-10a0b8082918-kube-api-access-nfw59" (OuterVolumeSpecName: "kube-api-access-nfw59") pod "5311a655-a897-486e-80ce-10a0b8082918" (UID: "5311a655-a897-486e-80ce-10a0b8082918"). InnerVolumeSpecName "kube-api-access-nfw59". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.133936 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5311a655-a897-486e-80ce-10a0b8082918-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5311a655-a897-486e-80ce-10a0b8082918" (UID: "5311a655-a897-486e-80ce-10a0b8082918"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.146194 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5311a655-a897-486e-80ce-10a0b8082918-inventory" (OuterVolumeSpecName: "inventory") pod "5311a655-a897-486e-80ce-10a0b8082918" (UID: "5311a655-a897-486e-80ce-10a0b8082918"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.209534 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfw59\" (UniqueName: \"kubernetes.io/projected/5311a655-a897-486e-80ce-10a0b8082918-kube-api-access-nfw59\") on node \"crc\" DevicePath \"\"" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.209561 4998 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5311a655-a897-486e-80ce-10a0b8082918-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.209571 4998 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5311a655-a897-486e-80ce-10a0b8082918-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.582381 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbwwm" event={"ID":"5311a655-a897-486e-80ce-10a0b8082918","Type":"ContainerDied","Data":"d89379d42acbfe0c7776fc385ecbae4281a7a441fe32332420631b0a119c9661"} Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.582777 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d89379d42acbfe0c7776fc385ecbae4281a7a441fe32332420631b0a119c9661" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.582437 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fbwwm" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.652838 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6"] Dec 03 16:30:18 crc kubenswrapper[4998]: E1203 16:30:18.653324 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5311a655-a897-486e-80ce-10a0b8082918" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.653343 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="5311a655-a897-486e-80ce-10a0b8082918" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.653566 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="5311a655-a897-486e-80ce-10a0b8082918" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.654248 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.656667 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.656733 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kcj9x" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.656806 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.656864 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.666000 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6"] Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.719697 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hm9lq\" (UniqueName: \"kubernetes.io/projected/e700335e-e6da-4bd2-ad00-6a7ecb9761a0-kube-api-access-hm9lq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6\" (UID: \"e700335e-e6da-4bd2-ad00-6a7ecb9761a0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.720023 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e700335e-e6da-4bd2-ad00-6a7ecb9761a0-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6\" (UID: \"e700335e-e6da-4bd2-ad00-6a7ecb9761a0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.720122 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e700335e-e6da-4bd2-ad00-6a7ecb9761a0-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6\" (UID: \"e700335e-e6da-4bd2-ad00-6a7ecb9761a0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.720246 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e700335e-e6da-4bd2-ad00-6a7ecb9761a0-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6\" (UID: \"e700335e-e6da-4bd2-ad00-6a7ecb9761a0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.822513 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e700335e-e6da-4bd2-ad00-6a7ecb9761a0-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6\" (UID: \"e700335e-e6da-4bd2-ad00-6a7ecb9761a0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.822593 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e700335e-e6da-4bd2-ad00-6a7ecb9761a0-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6\" (UID: \"e700335e-e6da-4bd2-ad00-6a7ecb9761a0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.822661 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e700335e-e6da-4bd2-ad00-6a7ecb9761a0-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6\" (UID: \"e700335e-e6da-4bd2-ad00-6a7ecb9761a0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.822718 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hm9lq\" (UniqueName: \"kubernetes.io/projected/e700335e-e6da-4bd2-ad00-6a7ecb9761a0-kube-api-access-hm9lq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6\" (UID: \"e700335e-e6da-4bd2-ad00-6a7ecb9761a0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.828654 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e700335e-e6da-4bd2-ad00-6a7ecb9761a0-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6\" (UID: \"e700335e-e6da-4bd2-ad00-6a7ecb9761a0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.829442 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e700335e-e6da-4bd2-ad00-6a7ecb9761a0-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6\" (UID: \"e700335e-e6da-4bd2-ad00-6a7ecb9761a0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.834965 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e700335e-e6da-4bd2-ad00-6a7ecb9761a0-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6\" (UID: \"e700335e-e6da-4bd2-ad00-6a7ecb9761a0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.839007 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hm9lq\" (UniqueName: \"kubernetes.io/projected/e700335e-e6da-4bd2-ad00-6a7ecb9761a0-kube-api-access-hm9lq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6\" (UID: \"e700335e-e6da-4bd2-ad00-6a7ecb9761a0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6" Dec 03 16:30:18 crc kubenswrapper[4998]: I1203 16:30:18.972837 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6" Dec 03 16:30:19 crc kubenswrapper[4998]: I1203 16:30:19.501711 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6"] Dec 03 16:30:19 crc kubenswrapper[4998]: I1203 16:30:19.592814 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6" event={"ID":"e700335e-e6da-4bd2-ad00-6a7ecb9761a0","Type":"ContainerStarted","Data":"8936a19fb821f7d589ed10409c1d2dac4d4a31730b5bfbc007803d50ad4275a8"} Dec 03 16:30:21 crc kubenswrapper[4998]: I1203 16:30:21.618332 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6" event={"ID":"e700335e-e6da-4bd2-ad00-6a7ecb9761a0","Type":"ContainerStarted","Data":"97f99c34183f734849f1740a0922d2559587a62319087f085cc62caadacc7d5e"} Dec 03 16:30:21 crc kubenswrapper[4998]: I1203 16:30:21.650250 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6" podStartSLOduration=2.7061786310000002 podStartE2EDuration="3.650232095s" podCreationTimestamp="2025-12-03 16:30:18 +0000 UTC" firstStartedPulling="2025-12-03 16:30:19.507698899 +0000 UTC m=+1598.119399122" lastFinishedPulling="2025-12-03 16:30:20.451752323 +0000 UTC m=+1599.063452586" observedRunningTime="2025-12-03 16:30:21.639180348 +0000 UTC m=+1600.250880581" watchObservedRunningTime="2025-12-03 16:30:21.650232095 +0000 UTC m=+1600.261932328" Dec 03 16:30:24 crc kubenswrapper[4998]: I1203 16:30:24.679965 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:30:24 crc kubenswrapper[4998]: E1203 16:30:24.681093 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:30:35 crc kubenswrapper[4998]: I1203 16:30:35.678271 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:30:35 crc kubenswrapper[4998]: E1203 16:30:35.679426 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:30:46 crc kubenswrapper[4998]: I1203 16:30:46.678869 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:30:46 crc kubenswrapper[4998]: E1203 16:30:46.679711 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:30:57 crc kubenswrapper[4998]: I1203 16:30:57.678277 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:30:57 crc kubenswrapper[4998]: E1203 16:30:57.679269 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:30:58 crc kubenswrapper[4998]: I1203 16:30:58.115733 4998 scope.go:117] "RemoveContainer" containerID="dce18792baccb8a737a1d6afcfdeae6a5b72eb876a08b9697f50acde4e2584ba" Dec 03 16:31:08 crc kubenswrapper[4998]: I1203 16:31:08.677611 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:31:08 crc kubenswrapper[4998]: E1203 16:31:08.678498 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:31:19 crc kubenswrapper[4998]: I1203 16:31:19.678723 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:31:19 crc kubenswrapper[4998]: E1203 16:31:19.679388 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:31:34 crc kubenswrapper[4998]: I1203 16:31:34.678613 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:31:34 crc kubenswrapper[4998]: E1203 16:31:34.679692 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:31:47 crc kubenswrapper[4998]: I1203 16:31:47.677489 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:31:47 crc kubenswrapper[4998]: E1203 16:31:47.678410 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:31:58 crc kubenswrapper[4998]: I1203 16:31:58.206718 4998 scope.go:117] "RemoveContainer" containerID="a1d30b67a1dfea698ecb9b1bea63314a38f9de7898fc55974c57c35c76092c39" Dec 03 16:31:58 crc kubenswrapper[4998]: I1203 16:31:58.246616 4998 scope.go:117] "RemoveContainer" containerID="5ab8347b81ea9f11541eee83ec6cd743c0c89c981fcef844dbf98fb32b6c3599" Dec 03 16:31:59 crc kubenswrapper[4998]: I1203 16:31:59.677906 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:31:59 crc kubenswrapper[4998]: E1203 16:31:59.678494 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:32:10 crc kubenswrapper[4998]: I1203 16:32:10.678042 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:32:10 crc kubenswrapper[4998]: E1203 16:32:10.679022 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:32:24 crc kubenswrapper[4998]: I1203 16:32:24.678453 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:32:24 crc kubenswrapper[4998]: E1203 16:32:24.679328 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:32:36 crc kubenswrapper[4998]: I1203 16:32:36.679660 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:32:36 crc kubenswrapper[4998]: E1203 16:32:36.681686 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:32:47 crc kubenswrapper[4998]: I1203 16:32:47.678605 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:32:47 crc kubenswrapper[4998]: E1203 16:32:47.679693 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:32:58 crc kubenswrapper[4998]: I1203 16:32:58.055897 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-21ed-account-create-update-qnv6c"] Dec 03 16:32:58 crc kubenswrapper[4998]: I1203 16:32:58.065950 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-21ed-account-create-update-qnv6c"] Dec 03 16:32:59 crc kubenswrapper[4998]: I1203 16:32:59.045249 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-fffa-account-create-update-bh4qr"] Dec 03 16:32:59 crc kubenswrapper[4998]: I1203 16:32:59.066375 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-p4448"] Dec 03 16:32:59 crc kubenswrapper[4998]: I1203 16:32:59.079386 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-db-create-r4skm"] Dec 03 16:32:59 crc kubenswrapper[4998]: I1203 16:32:59.088024 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-db-create-r4skm"] Dec 03 16:32:59 crc kubenswrapper[4998]: I1203 16:32:59.095877 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-fffa-account-create-update-bh4qr"] Dec 03 16:32:59 crc kubenswrapper[4998]: I1203 16:32:59.104809 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-p4448"] Dec 03 16:32:59 crc kubenswrapper[4998]: I1203 16:32:59.114120 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-gwldb"] Dec 03 16:32:59 crc kubenswrapper[4998]: I1203 16:32:59.123080 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-gwldb"] Dec 03 16:32:59 crc kubenswrapper[4998]: I1203 16:32:59.133355 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-daeb-account-create-update-vsrjb"] Dec 03 16:32:59 crc kubenswrapper[4998]: I1203 16:32:59.141899 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-daeb-account-create-update-vsrjb"] Dec 03 16:32:59 crc kubenswrapper[4998]: I1203 16:32:59.691927 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ce04863-7333-4484-9507-93143d276e47" path="/var/lib/kubelet/pods/0ce04863-7333-4484-9507-93143d276e47/volumes" Dec 03 16:32:59 crc kubenswrapper[4998]: I1203 16:32:59.692930 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15941ab6-753a-4eb7-8e6f-51ffae3a0227" path="/var/lib/kubelet/pods/15941ab6-753a-4eb7-8e6f-51ffae3a0227/volumes" Dec 03 16:32:59 crc kubenswrapper[4998]: I1203 16:32:59.693923 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3739b197-1e59-48fa-a7f7-9e73b95a382a" path="/var/lib/kubelet/pods/3739b197-1e59-48fa-a7f7-9e73b95a382a/volumes" Dec 03 16:32:59 crc kubenswrapper[4998]: I1203 16:32:59.695190 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72129c14-7e4f-4e0a-9053-7d2178a835ae" path="/var/lib/kubelet/pods/72129c14-7e4f-4e0a-9053-7d2178a835ae/volumes" Dec 03 16:32:59 crc kubenswrapper[4998]: I1203 16:32:59.697074 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b542f7b1-ca06-4b56-b45c-dccde2fca019" path="/var/lib/kubelet/pods/b542f7b1-ca06-4b56-b45c-dccde2fca019/volumes" Dec 03 16:32:59 crc kubenswrapper[4998]: I1203 16:32:59.697952 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d71c5697-f3d2-4c29-a722-49823fa55d6d" path="/var/lib/kubelet/pods/d71c5697-f3d2-4c29-a722-49823fa55d6d/volumes" Dec 03 16:33:01 crc kubenswrapper[4998]: I1203 16:33:01.685011 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:33:01 crc kubenswrapper[4998]: E1203 16:33:01.685566 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:33:14 crc kubenswrapper[4998]: I1203 16:33:14.678582 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:33:14 crc kubenswrapper[4998]: E1203 16:33:14.679593 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:33:29 crc kubenswrapper[4998]: I1203 16:33:29.678927 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:33:30 crc kubenswrapper[4998]: I1203 16:33:30.917492 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"9001df2bc0dc33071d80330946bfdea9e78811028533587a680aea91a458fee6"} Dec 03 16:33:31 crc kubenswrapper[4998]: I1203 16:33:31.050582 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-b4hsk"] Dec 03 16:33:31 crc kubenswrapper[4998]: I1203 16:33:31.063739 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-rsc5q"] Dec 03 16:33:31 crc kubenswrapper[4998]: I1203 16:33:31.077272 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-f9d0-account-create-update-pf87v"] Dec 03 16:33:31 crc kubenswrapper[4998]: I1203 16:33:31.087051 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-cm9s2"] Dec 03 16:33:31 crc kubenswrapper[4998]: I1203 16:33:31.095836 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-zv5gc"] Dec 03 16:33:31 crc kubenswrapper[4998]: I1203 16:33:31.103622 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-7a2f-account-create-update-fjb4t"] Dec 03 16:33:31 crc kubenswrapper[4998]: I1203 16:33:31.112252 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-b4hsk"] Dec 03 16:33:31 crc kubenswrapper[4998]: I1203 16:33:31.120157 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-zv5gc"] Dec 03 16:33:31 crc kubenswrapper[4998]: I1203 16:33:31.129048 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-rsc5q"] Dec 03 16:33:31 crc kubenswrapper[4998]: I1203 16:33:31.138181 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-f9d0-account-create-update-pf87v"] Dec 03 16:33:31 crc kubenswrapper[4998]: I1203 16:33:31.148113 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-7a2f-account-create-update-fjb4t"] Dec 03 16:33:31 crc kubenswrapper[4998]: I1203 16:33:31.156064 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-cm9s2"] Dec 03 16:33:31 crc kubenswrapper[4998]: I1203 16:33:31.163188 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-ea32-account-create-update-zszrs"] Dec 03 16:33:31 crc kubenswrapper[4998]: I1203 16:33:31.171032 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-7f14-account-create-update-kb9kq"] Dec 03 16:33:31 crc kubenswrapper[4998]: I1203 16:33:31.179263 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-ea32-account-create-update-zszrs"] Dec 03 16:33:31 crc kubenswrapper[4998]: I1203 16:33:31.186943 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-7f14-account-create-update-kb9kq"] Dec 03 16:33:31 crc kubenswrapper[4998]: I1203 16:33:31.691569 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00a5807c-76d9-4a03-b8f1-1c8823915d32" path="/var/lib/kubelet/pods/00a5807c-76d9-4a03-b8f1-1c8823915d32/volumes" Dec 03 16:33:31 crc kubenswrapper[4998]: I1203 16:33:31.693270 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20531474-cb0d-4ce8-90fc-e9974acc6d1a" path="/var/lib/kubelet/pods/20531474-cb0d-4ce8-90fc-e9974acc6d1a/volumes" Dec 03 16:33:31 crc kubenswrapper[4998]: I1203 16:33:31.694309 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23518b26-e4ce-495d-aada-91f784244125" path="/var/lib/kubelet/pods/23518b26-e4ce-495d-aada-91f784244125/volumes" Dec 03 16:33:31 crc kubenswrapper[4998]: I1203 16:33:31.695655 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a4bd716-a01d-4e69-90ee-ae27b0e77c85" path="/var/lib/kubelet/pods/2a4bd716-a01d-4e69-90ee-ae27b0e77c85/volumes" Dec 03 16:33:31 crc kubenswrapper[4998]: I1203 16:33:31.697302 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cb17c41-9a33-45df-b013-044e94f413b4" path="/var/lib/kubelet/pods/8cb17c41-9a33-45df-b013-044e94f413b4/volumes" Dec 03 16:33:31 crc kubenswrapper[4998]: I1203 16:33:31.698451 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fc53f31-60a9-4a2d-9d61-6c3263f95ca9" path="/var/lib/kubelet/pods/8fc53f31-60a9-4a2d-9d61-6c3263f95ca9/volumes" Dec 03 16:33:31 crc kubenswrapper[4998]: I1203 16:33:31.699191 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8c2151f-9cdb-445e-865c-572da28ab46c" path="/var/lib/kubelet/pods/c8c2151f-9cdb-445e-865c-572da28ab46c/volumes" Dec 03 16:33:31 crc kubenswrapper[4998]: I1203 16:33:31.700857 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fecdd245-3377-4b65-b04d-50da92ee757b" path="/var/lib/kubelet/pods/fecdd245-3377-4b65-b04d-50da92ee757b/volumes" Dec 03 16:33:51 crc kubenswrapper[4998]: I1203 16:33:51.142531 4998 generic.go:334] "Generic (PLEG): container finished" podID="e700335e-e6da-4bd2-ad00-6a7ecb9761a0" containerID="97f99c34183f734849f1740a0922d2559587a62319087f085cc62caadacc7d5e" exitCode=0 Dec 03 16:33:51 crc kubenswrapper[4998]: I1203 16:33:51.142602 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6" event={"ID":"e700335e-e6da-4bd2-ad00-6a7ecb9761a0","Type":"ContainerDied","Data":"97f99c34183f734849f1740a0922d2559587a62319087f085cc62caadacc7d5e"} Dec 03 16:33:52 crc kubenswrapper[4998]: I1203 16:33:52.676673 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6" Dec 03 16:33:52 crc kubenswrapper[4998]: I1203 16:33:52.711414 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hm9lq\" (UniqueName: \"kubernetes.io/projected/e700335e-e6da-4bd2-ad00-6a7ecb9761a0-kube-api-access-hm9lq\") pod \"e700335e-e6da-4bd2-ad00-6a7ecb9761a0\" (UID: \"e700335e-e6da-4bd2-ad00-6a7ecb9761a0\") " Dec 03 16:33:52 crc kubenswrapper[4998]: I1203 16:33:52.711595 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e700335e-e6da-4bd2-ad00-6a7ecb9761a0-ssh-key\") pod \"e700335e-e6da-4bd2-ad00-6a7ecb9761a0\" (UID: \"e700335e-e6da-4bd2-ad00-6a7ecb9761a0\") " Dec 03 16:33:52 crc kubenswrapper[4998]: I1203 16:33:52.711651 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e700335e-e6da-4bd2-ad00-6a7ecb9761a0-bootstrap-combined-ca-bundle\") pod \"e700335e-e6da-4bd2-ad00-6a7ecb9761a0\" (UID: \"e700335e-e6da-4bd2-ad00-6a7ecb9761a0\") " Dec 03 16:33:52 crc kubenswrapper[4998]: I1203 16:33:52.711771 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e700335e-e6da-4bd2-ad00-6a7ecb9761a0-inventory\") pod \"e700335e-e6da-4bd2-ad00-6a7ecb9761a0\" (UID: \"e700335e-e6da-4bd2-ad00-6a7ecb9761a0\") " Dec 03 16:33:52 crc kubenswrapper[4998]: I1203 16:33:52.720464 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e700335e-e6da-4bd2-ad00-6a7ecb9761a0-kube-api-access-hm9lq" (OuterVolumeSpecName: "kube-api-access-hm9lq") pod "e700335e-e6da-4bd2-ad00-6a7ecb9761a0" (UID: "e700335e-e6da-4bd2-ad00-6a7ecb9761a0"). InnerVolumeSpecName "kube-api-access-hm9lq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:33:52 crc kubenswrapper[4998]: I1203 16:33:52.723742 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e700335e-e6da-4bd2-ad00-6a7ecb9761a0-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "e700335e-e6da-4bd2-ad00-6a7ecb9761a0" (UID: "e700335e-e6da-4bd2-ad00-6a7ecb9761a0"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:33:52 crc kubenswrapper[4998]: I1203 16:33:52.758266 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e700335e-e6da-4bd2-ad00-6a7ecb9761a0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e700335e-e6da-4bd2-ad00-6a7ecb9761a0" (UID: "e700335e-e6da-4bd2-ad00-6a7ecb9761a0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:33:52 crc kubenswrapper[4998]: I1203 16:33:52.760201 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e700335e-e6da-4bd2-ad00-6a7ecb9761a0-inventory" (OuterVolumeSpecName: "inventory") pod "e700335e-e6da-4bd2-ad00-6a7ecb9761a0" (UID: "e700335e-e6da-4bd2-ad00-6a7ecb9761a0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:33:52 crc kubenswrapper[4998]: I1203 16:33:52.814231 4998 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e700335e-e6da-4bd2-ad00-6a7ecb9761a0-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 16:33:52 crc kubenswrapper[4998]: I1203 16:33:52.814264 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hm9lq\" (UniqueName: \"kubernetes.io/projected/e700335e-e6da-4bd2-ad00-6a7ecb9761a0-kube-api-access-hm9lq\") on node \"crc\" DevicePath \"\"" Dec 03 16:33:52 crc kubenswrapper[4998]: I1203 16:33:52.814274 4998 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e700335e-e6da-4bd2-ad00-6a7ecb9761a0-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 16:33:52 crc kubenswrapper[4998]: I1203 16:33:52.814286 4998 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e700335e-e6da-4bd2-ad00-6a7ecb9761a0-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:33:53 crc kubenswrapper[4998]: I1203 16:33:53.169843 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6" event={"ID":"e700335e-e6da-4bd2-ad00-6a7ecb9761a0","Type":"ContainerDied","Data":"8936a19fb821f7d589ed10409c1d2dac4d4a31730b5bfbc007803d50ad4275a8"} Dec 03 16:33:53 crc kubenswrapper[4998]: I1203 16:33:53.169908 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8936a19fb821f7d589ed10409c1d2dac4d4a31730b5bfbc007803d50ad4275a8" Dec 03 16:33:53 crc kubenswrapper[4998]: I1203 16:33:53.169869 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6" Dec 03 16:33:53 crc kubenswrapper[4998]: I1203 16:33:53.282973 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q"] Dec 03 16:33:53 crc kubenswrapper[4998]: E1203 16:33:53.283612 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e700335e-e6da-4bd2-ad00-6a7ecb9761a0" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 16:33:53 crc kubenswrapper[4998]: I1203 16:33:53.283632 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="e700335e-e6da-4bd2-ad00-6a7ecb9761a0" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 16:33:53 crc kubenswrapper[4998]: I1203 16:33:53.284522 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="e700335e-e6da-4bd2-ad00-6a7ecb9761a0" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 16:33:53 crc kubenswrapper[4998]: I1203 16:33:53.285206 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q" Dec 03 16:33:53 crc kubenswrapper[4998]: I1203 16:33:53.286971 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kcj9x" Dec 03 16:33:53 crc kubenswrapper[4998]: I1203 16:33:53.288319 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 16:33:53 crc kubenswrapper[4998]: I1203 16:33:53.291021 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 16:33:53 crc kubenswrapper[4998]: I1203 16:33:53.291894 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 16:33:53 crc kubenswrapper[4998]: I1203 16:33:53.294501 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q"] Dec 03 16:33:53 crc kubenswrapper[4998]: I1203 16:33:53.324235 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/742311fa-8f91-4738-ae24-66e591d9b8a1-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q\" (UID: \"742311fa-8f91-4738-ae24-66e591d9b8a1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q" Dec 03 16:33:53 crc kubenswrapper[4998]: I1203 16:33:53.324348 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6c68\" (UniqueName: \"kubernetes.io/projected/742311fa-8f91-4738-ae24-66e591d9b8a1-kube-api-access-b6c68\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q\" (UID: \"742311fa-8f91-4738-ae24-66e591d9b8a1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q" Dec 03 16:33:53 crc kubenswrapper[4998]: I1203 16:33:53.324397 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/742311fa-8f91-4738-ae24-66e591d9b8a1-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q\" (UID: \"742311fa-8f91-4738-ae24-66e591d9b8a1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q" Dec 03 16:33:53 crc kubenswrapper[4998]: I1203 16:33:53.425981 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/742311fa-8f91-4738-ae24-66e591d9b8a1-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q\" (UID: \"742311fa-8f91-4738-ae24-66e591d9b8a1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q" Dec 03 16:33:53 crc kubenswrapper[4998]: I1203 16:33:53.426076 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6c68\" (UniqueName: \"kubernetes.io/projected/742311fa-8f91-4738-ae24-66e591d9b8a1-kube-api-access-b6c68\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q\" (UID: \"742311fa-8f91-4738-ae24-66e591d9b8a1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q" Dec 03 16:33:53 crc kubenswrapper[4998]: I1203 16:33:53.426116 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/742311fa-8f91-4738-ae24-66e591d9b8a1-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q\" (UID: \"742311fa-8f91-4738-ae24-66e591d9b8a1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q" Dec 03 16:33:53 crc kubenswrapper[4998]: I1203 16:33:53.431654 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/742311fa-8f91-4738-ae24-66e591d9b8a1-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q\" (UID: \"742311fa-8f91-4738-ae24-66e591d9b8a1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q" Dec 03 16:33:53 crc kubenswrapper[4998]: I1203 16:33:53.440861 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/742311fa-8f91-4738-ae24-66e591d9b8a1-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q\" (UID: \"742311fa-8f91-4738-ae24-66e591d9b8a1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q" Dec 03 16:33:53 crc kubenswrapper[4998]: I1203 16:33:53.443339 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6c68\" (UniqueName: \"kubernetes.io/projected/742311fa-8f91-4738-ae24-66e591d9b8a1-kube-api-access-b6c68\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q\" (UID: \"742311fa-8f91-4738-ae24-66e591d9b8a1\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q" Dec 03 16:33:53 crc kubenswrapper[4998]: I1203 16:33:53.607648 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q" Dec 03 16:33:55 crc kubenswrapper[4998]: I1203 16:33:55.084089 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q"] Dec 03 16:33:55 crc kubenswrapper[4998]: I1203 16:33:55.630631 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q" event={"ID":"742311fa-8f91-4738-ae24-66e591d9b8a1","Type":"ContainerStarted","Data":"6e29f7e69a77627695fc2394d8dbe23415068406929314ce5789560bd072fdf0"} Dec 03 16:33:56 crc kubenswrapper[4998]: I1203 16:33:56.050353 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-nj4w6"] Dec 03 16:33:56 crc kubenswrapper[4998]: I1203 16:33:56.065217 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-db-sync-rjhmp"] Dec 03 16:33:56 crc kubenswrapper[4998]: I1203 16:33:56.075816 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-nj4w6"] Dec 03 16:33:56 crc kubenswrapper[4998]: I1203 16:33:56.085078 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-db-sync-rjhmp"] Dec 03 16:33:56 crc kubenswrapper[4998]: I1203 16:33:56.643216 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q" event={"ID":"742311fa-8f91-4738-ae24-66e591d9b8a1","Type":"ContainerStarted","Data":"37af099517bdd319cebc2c10d9aa950f7734fceed3ca34801eac1eec50ac3b52"} Dec 03 16:33:56 crc kubenswrapper[4998]: I1203 16:33:56.665909 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q" podStartSLOduration=2.8007509390000003 podStartE2EDuration="3.665892027s" podCreationTimestamp="2025-12-03 16:33:53 +0000 UTC" firstStartedPulling="2025-12-03 16:33:55.095539051 +0000 UTC m=+1813.707239274" lastFinishedPulling="2025-12-03 16:33:55.960680129 +0000 UTC m=+1814.572380362" observedRunningTime="2025-12-03 16:33:56.660351109 +0000 UTC m=+1815.272051342" watchObservedRunningTime="2025-12-03 16:33:56.665892027 +0000 UTC m=+1815.277592250" Dec 03 16:33:57 crc kubenswrapper[4998]: I1203 16:33:57.693343 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e11bf44e-799b-472b-aafb-89d5d071ad91" path="/var/lib/kubelet/pods/e11bf44e-799b-472b-aafb-89d5d071ad91/volumes" Dec 03 16:33:57 crc kubenswrapper[4998]: I1203 16:33:57.694596 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edc74770-171e-4ade-8dcd-bcaead186741" path="/var/lib/kubelet/pods/edc74770-171e-4ade-8dcd-bcaead186741/volumes" Dec 03 16:33:58 crc kubenswrapper[4998]: I1203 16:33:58.366814 4998 scope.go:117] "RemoveContainer" containerID="4140050d1ad6858044ba4f30d05c0f46fdc7befefd812effe581a96fa4ad4b18" Dec 03 16:33:58 crc kubenswrapper[4998]: I1203 16:33:58.412511 4998 scope.go:117] "RemoveContainer" containerID="fedfa179097767efc0d2374a3a66b294f4cc1edfacd2dd7f9eb72f4f28be735a" Dec 03 16:33:58 crc kubenswrapper[4998]: I1203 16:33:58.443841 4998 scope.go:117] "RemoveContainer" containerID="9184e20036fae6dea7fb09ef9673244bd0545bc4d1b9ef053b11aced17e6fe51" Dec 03 16:33:58 crc kubenswrapper[4998]: I1203 16:33:58.494464 4998 scope.go:117] "RemoveContainer" containerID="66b40d6d558616b9828bc83b8a9c16287350b7bba9b9fcc34de50c5dfa9e4466" Dec 03 16:33:58 crc kubenswrapper[4998]: I1203 16:33:58.521048 4998 scope.go:117] "RemoveContainer" containerID="1d2cb0ab5309267aa896af5145b7f88c89ffdf464eaa0fe4029fd43bd43ed634" Dec 03 16:33:58 crc kubenswrapper[4998]: I1203 16:33:58.577966 4998 scope.go:117] "RemoveContainer" containerID="f5cf4c1c2b2ded19a3dd250a7981ea172d494c29024b0caffeef55b427f76d43" Dec 03 16:33:58 crc kubenswrapper[4998]: I1203 16:33:58.598653 4998 scope.go:117] "RemoveContainer" containerID="c6719e5a27b5be3a5194c080d34db81c4c39ac724df7ddabfa3d9de6ce711dc9" Dec 03 16:33:58 crc kubenswrapper[4998]: I1203 16:33:58.645861 4998 scope.go:117] "RemoveContainer" containerID="9fe611c4100838470d9ddd83b84950c87202cecce825b8bc1f19e188b93071e6" Dec 03 16:33:58 crc kubenswrapper[4998]: I1203 16:33:58.669990 4998 scope.go:117] "RemoveContainer" containerID="50b57c87c0146a67455069dc0f21e106f57997c31cb455ec6bab3e1ee1d4a159" Dec 03 16:33:58 crc kubenswrapper[4998]: I1203 16:33:58.721943 4998 scope.go:117] "RemoveContainer" containerID="e8281d3840790c2c034b305a7252b0b14cc1498567f5d7cb52a02564cb924e6d" Dec 03 16:33:58 crc kubenswrapper[4998]: I1203 16:33:58.752855 4998 scope.go:117] "RemoveContainer" containerID="d95b944c60c5729c15567f7c9cd0d472ca7f9356c0a1849a2e87b6e9cef89446" Dec 03 16:33:58 crc kubenswrapper[4998]: I1203 16:33:58.774936 4998 scope.go:117] "RemoveContainer" containerID="88950ec72a9b62516073acb253b694726d83368ce8222ef79a223454a9bba802" Dec 03 16:33:58 crc kubenswrapper[4998]: I1203 16:33:58.808297 4998 scope.go:117] "RemoveContainer" containerID="4f14427fc6d32bd017613ec2b7d3009f7f7eab8658a1964b632c6af7dbc42fde" Dec 03 16:33:58 crc kubenswrapper[4998]: I1203 16:33:58.910199 4998 scope.go:117] "RemoveContainer" containerID="32c9f837e40269fa018e70f76d3c57e75cf93e48838768e46b94217965185dbb" Dec 03 16:33:58 crc kubenswrapper[4998]: I1203 16:33:58.969123 4998 scope.go:117] "RemoveContainer" containerID="1799f2ac959a3ee61295e4179dcd841540aad9731c9842955db0dbccd29b29b9" Dec 03 16:33:58 crc kubenswrapper[4998]: I1203 16:33:58.991674 4998 scope.go:117] "RemoveContainer" containerID="75385576e42fd8cb783f5ce0d82d0c0520f8d6a3e21af41b30c90d912cd5f3b0" Dec 03 16:33:59 crc kubenswrapper[4998]: I1203 16:33:59.019058 4998 scope.go:117] "RemoveContainer" containerID="02b6c984776989d01a404c0876ed79a235d0feb077b46a3356310c2a94dffa78" Dec 03 16:33:59 crc kubenswrapper[4998]: I1203 16:33:59.039095 4998 scope.go:117] "RemoveContainer" containerID="2f36e3df4d17f0f527d02f6c2f1f136d16d80742bb0abcc3a1ebc80c76a50420" Dec 03 16:33:59 crc kubenswrapper[4998]: I1203 16:33:59.060625 4998 scope.go:117] "RemoveContainer" containerID="854db052dba6cd5655d3c5d1c22c7dd1fd27a486305b18052c7f7d315ddf755d" Dec 03 16:33:59 crc kubenswrapper[4998]: I1203 16:33:59.079348 4998 scope.go:117] "RemoveContainer" containerID="dcc8c7984741634c34bd1ede60a7504fc741bda719000c93a8ff3754c2e0c699" Dec 03 16:34:46 crc kubenswrapper[4998]: I1203 16:34:46.051133 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-bf2md"] Dec 03 16:34:46 crc kubenswrapper[4998]: I1203 16:34:46.059354 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-bf2md"] Dec 03 16:34:47 crc kubenswrapper[4998]: I1203 16:34:47.689816 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="751f7a77-ec1d-4529-8762-a48fe772c25b" path="/var/lib/kubelet/pods/751f7a77-ec1d-4529-8762-a48fe772c25b/volumes" Dec 03 16:34:59 crc kubenswrapper[4998]: I1203 16:34:59.384396 4998 scope.go:117] "RemoveContainer" containerID="3e0b15332e4f017b612e36e98b7d4c7901623e0813d95832b81628ee8a6bbaeb" Dec 03 16:35:00 crc kubenswrapper[4998]: I1203 16:35:00.068507 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-qw92l"] Dec 03 16:35:00 crc kubenswrapper[4998]: I1203 16:35:00.082515 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-qw92l"] Dec 03 16:35:01 crc kubenswrapper[4998]: I1203 16:35:01.695246 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cca43ea-2244-4f5c-8704-8d91634c0ad6" path="/var/lib/kubelet/pods/5cca43ea-2244-4f5c-8704-8d91634c0ad6/volumes" Dec 03 16:35:08 crc kubenswrapper[4998]: I1203 16:35:08.032335 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-w8k5f"] Dec 03 16:35:08 crc kubenswrapper[4998]: I1203 16:35:08.040503 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-w8k5f"] Dec 03 16:35:09 crc kubenswrapper[4998]: I1203 16:35:09.699677 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d2429d8-9baa-4532-9f8b-9ffb72a3a764" path="/var/lib/kubelet/pods/5d2429d8-9baa-4532-9f8b-9ffb72a3a764/volumes" Dec 03 16:35:19 crc kubenswrapper[4998]: I1203 16:35:19.036440 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-8lz9j"] Dec 03 16:35:19 crc kubenswrapper[4998]: I1203 16:35:19.046164 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-8lz9j"] Dec 03 16:35:19 crc kubenswrapper[4998]: I1203 16:35:19.692632 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82aa9d23-457c-45a0-9729-e5a803738237" path="/var/lib/kubelet/pods/82aa9d23-457c-45a0-9729-e5a803738237/volumes" Dec 03 16:35:27 crc kubenswrapper[4998]: I1203 16:35:27.035158 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-2g9pt"] Dec 03 16:35:27 crc kubenswrapper[4998]: I1203 16:35:27.046392 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-2g9pt"] Dec 03 16:35:27 crc kubenswrapper[4998]: I1203 16:35:27.689379 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac00cda5-18ab-48de-a82e-1ebe08fa518c" path="/var/lib/kubelet/pods/ac00cda5-18ab-48de-a82e-1ebe08fa518c/volumes" Dec 03 16:35:29 crc kubenswrapper[4998]: I1203 16:35:29.029580 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-2rqhm"] Dec 03 16:35:29 crc kubenswrapper[4998]: I1203 16:35:29.039629 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-2rqhm"] Dec 03 16:35:29 crc kubenswrapper[4998]: I1203 16:35:29.692535 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de77cff6-7039-418c-8114-3c64ea4ce8e4" path="/var/lib/kubelet/pods/de77cff6-7039-418c-8114-3c64ea4ce8e4/volumes" Dec 03 16:35:50 crc kubenswrapper[4998]: I1203 16:35:50.074961 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-667sw"] Dec 03 16:35:50 crc kubenswrapper[4998]: I1203 16:35:50.078364 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-667sw" Dec 03 16:35:50 crc kubenswrapper[4998]: I1203 16:35:50.096891 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-667sw"] Dec 03 16:35:50 crc kubenswrapper[4998]: I1203 16:35:50.217906 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc03c421-4ea1-46c9-ae67-03af108bf717-utilities\") pod \"redhat-marketplace-667sw\" (UID: \"fc03c421-4ea1-46c9-ae67-03af108bf717\") " pod="openshift-marketplace/redhat-marketplace-667sw" Dec 03 16:35:50 crc kubenswrapper[4998]: I1203 16:35:50.218099 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc03c421-4ea1-46c9-ae67-03af108bf717-catalog-content\") pod \"redhat-marketplace-667sw\" (UID: \"fc03c421-4ea1-46c9-ae67-03af108bf717\") " pod="openshift-marketplace/redhat-marketplace-667sw" Dec 03 16:35:50 crc kubenswrapper[4998]: I1203 16:35:50.218302 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6lzd\" (UniqueName: \"kubernetes.io/projected/fc03c421-4ea1-46c9-ae67-03af108bf717-kube-api-access-d6lzd\") pod \"redhat-marketplace-667sw\" (UID: \"fc03c421-4ea1-46c9-ae67-03af108bf717\") " pod="openshift-marketplace/redhat-marketplace-667sw" Dec 03 16:35:50 crc kubenswrapper[4998]: I1203 16:35:50.320119 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc03c421-4ea1-46c9-ae67-03af108bf717-utilities\") pod \"redhat-marketplace-667sw\" (UID: \"fc03c421-4ea1-46c9-ae67-03af108bf717\") " pod="openshift-marketplace/redhat-marketplace-667sw" Dec 03 16:35:50 crc kubenswrapper[4998]: I1203 16:35:50.320184 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc03c421-4ea1-46c9-ae67-03af108bf717-catalog-content\") pod \"redhat-marketplace-667sw\" (UID: \"fc03c421-4ea1-46c9-ae67-03af108bf717\") " pod="openshift-marketplace/redhat-marketplace-667sw" Dec 03 16:35:50 crc kubenswrapper[4998]: I1203 16:35:50.320231 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6lzd\" (UniqueName: \"kubernetes.io/projected/fc03c421-4ea1-46c9-ae67-03af108bf717-kube-api-access-d6lzd\") pod \"redhat-marketplace-667sw\" (UID: \"fc03c421-4ea1-46c9-ae67-03af108bf717\") " pod="openshift-marketplace/redhat-marketplace-667sw" Dec 03 16:35:50 crc kubenswrapper[4998]: I1203 16:35:50.320739 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc03c421-4ea1-46c9-ae67-03af108bf717-utilities\") pod \"redhat-marketplace-667sw\" (UID: \"fc03c421-4ea1-46c9-ae67-03af108bf717\") " pod="openshift-marketplace/redhat-marketplace-667sw" Dec 03 16:35:50 crc kubenswrapper[4998]: I1203 16:35:50.320851 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc03c421-4ea1-46c9-ae67-03af108bf717-catalog-content\") pod \"redhat-marketplace-667sw\" (UID: \"fc03c421-4ea1-46c9-ae67-03af108bf717\") " pod="openshift-marketplace/redhat-marketplace-667sw" Dec 03 16:35:50 crc kubenswrapper[4998]: I1203 16:35:50.344230 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6lzd\" (UniqueName: \"kubernetes.io/projected/fc03c421-4ea1-46c9-ae67-03af108bf717-kube-api-access-d6lzd\") pod \"redhat-marketplace-667sw\" (UID: \"fc03c421-4ea1-46c9-ae67-03af108bf717\") " pod="openshift-marketplace/redhat-marketplace-667sw" Dec 03 16:35:50 crc kubenswrapper[4998]: I1203 16:35:50.406247 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-667sw" Dec 03 16:35:50 crc kubenswrapper[4998]: I1203 16:35:50.934204 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-667sw"] Dec 03 16:35:50 crc kubenswrapper[4998]: W1203 16:35:50.940749 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc03c421_4ea1_46c9_ae67_03af108bf717.slice/crio-af6719fbb0277d3806d982b2b721909c6a5d79fb0b270c178f2347de91651ae2 WatchSource:0}: Error finding container af6719fbb0277d3806d982b2b721909c6a5d79fb0b270c178f2347de91651ae2: Status 404 returned error can't find the container with id af6719fbb0277d3806d982b2b721909c6a5d79fb0b270c178f2347de91651ae2 Dec 03 16:35:51 crc kubenswrapper[4998]: I1203 16:35:51.919098 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-667sw" event={"ID":"fc03c421-4ea1-46c9-ae67-03af108bf717","Type":"ContainerStarted","Data":"af6719fbb0277d3806d982b2b721909c6a5d79fb0b270c178f2347de91651ae2"} Dec 03 16:35:52 crc kubenswrapper[4998]: I1203 16:35:52.932061 4998 generic.go:334] "Generic (PLEG): container finished" podID="fc03c421-4ea1-46c9-ae67-03af108bf717" containerID="010d3d4ab8f88b11b226c20f8044c10b3cd9cce526bce6084871fddd298b52c4" exitCode=0 Dec 03 16:35:52 crc kubenswrapper[4998]: I1203 16:35:52.932107 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-667sw" event={"ID":"fc03c421-4ea1-46c9-ae67-03af108bf717","Type":"ContainerDied","Data":"010d3d4ab8f88b11b226c20f8044c10b3cd9cce526bce6084871fddd298b52c4"} Dec 03 16:35:52 crc kubenswrapper[4998]: I1203 16:35:52.938155 4998 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 16:35:56 crc kubenswrapper[4998]: I1203 16:35:56.970869 4998 generic.go:334] "Generic (PLEG): container finished" podID="fc03c421-4ea1-46c9-ae67-03af108bf717" containerID="eb781609f60403f61d92e854e2c3566c29913d5b2205b97365077cc9d21bca07" exitCode=0 Dec 03 16:35:56 crc kubenswrapper[4998]: I1203 16:35:56.970965 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-667sw" event={"ID":"fc03c421-4ea1-46c9-ae67-03af108bf717","Type":"ContainerDied","Data":"eb781609f60403f61d92e854e2c3566c29913d5b2205b97365077cc9d21bca07"} Dec 03 16:35:57 crc kubenswrapper[4998]: I1203 16:35:57.110963 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:35:57 crc kubenswrapper[4998]: I1203 16:35:57.111039 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:35:58 crc kubenswrapper[4998]: I1203 16:35:58.989593 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-667sw" event={"ID":"fc03c421-4ea1-46c9-ae67-03af108bf717","Type":"ContainerStarted","Data":"7c35ec5d9e0f8d3c60f967ceac7eb7e26613ef6c4899ca9f7f501556135571b6"} Dec 03 16:35:59 crc kubenswrapper[4998]: I1203 16:35:59.015663 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-667sw" podStartSLOduration=3.747884922 podStartE2EDuration="9.015641172s" podCreationTimestamp="2025-12-03 16:35:50 +0000 UTC" firstStartedPulling="2025-12-03 16:35:52.93751061 +0000 UTC m=+1931.549210873" lastFinishedPulling="2025-12-03 16:35:58.2052669 +0000 UTC m=+1936.816967123" observedRunningTime="2025-12-03 16:35:59.007649083 +0000 UTC m=+1937.619349306" watchObservedRunningTime="2025-12-03 16:35:59.015641172 +0000 UTC m=+1937.627341395" Dec 03 16:35:59 crc kubenswrapper[4998]: I1203 16:35:59.466097 4998 scope.go:117] "RemoveContainer" containerID="2298bd80e5bdaea4315dbb4b6a0effa05cbdcae8e440ead4f9287b4f7bd6ed75" Dec 03 16:35:59 crc kubenswrapper[4998]: I1203 16:35:59.518996 4998 scope.go:117] "RemoveContainer" containerID="9c2779da866bb38d647d67f9337f9006e0141dc294f2b196b28cf1e700822fcf" Dec 03 16:35:59 crc kubenswrapper[4998]: I1203 16:35:59.549367 4998 scope.go:117] "RemoveContainer" containerID="d9de792d4b5c1b7f62defd70fa80c00e1a269c701e996345d575192a2ef656f2" Dec 03 16:35:59 crc kubenswrapper[4998]: I1203 16:35:59.613440 4998 scope.go:117] "RemoveContainer" containerID="a2dafe752937450125658f1ce9546da3c4d740014f80b43ed2fba0e66e8931f8" Dec 03 16:35:59 crc kubenswrapper[4998]: I1203 16:35:59.658533 4998 scope.go:117] "RemoveContainer" containerID="043af9beddc6495b4dd97ae2a362ea5bea94ef2a3d8e62d695b5f5765c933eda" Dec 03 16:36:00 crc kubenswrapper[4998]: I1203 16:36:00.406535 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-667sw" Dec 03 16:36:00 crc kubenswrapper[4998]: I1203 16:36:00.406960 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-667sw" Dec 03 16:36:00 crc kubenswrapper[4998]: I1203 16:36:00.460692 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-667sw" Dec 03 16:36:01 crc kubenswrapper[4998]: I1203 16:36:01.740632 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j6p6n"] Dec 03 16:36:01 crc kubenswrapper[4998]: I1203 16:36:01.743402 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j6p6n" Dec 03 16:36:01 crc kubenswrapper[4998]: I1203 16:36:01.756148 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j6p6n"] Dec 03 16:36:01 crc kubenswrapper[4998]: I1203 16:36:01.855363 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8833f82b-b833-4a08-864f-219787f8f4b8-catalog-content\") pod \"redhat-operators-j6p6n\" (UID: \"8833f82b-b833-4a08-864f-219787f8f4b8\") " pod="openshift-marketplace/redhat-operators-j6p6n" Dec 03 16:36:01 crc kubenswrapper[4998]: I1203 16:36:01.855446 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cmw4\" (UniqueName: \"kubernetes.io/projected/8833f82b-b833-4a08-864f-219787f8f4b8-kube-api-access-5cmw4\") pod \"redhat-operators-j6p6n\" (UID: \"8833f82b-b833-4a08-864f-219787f8f4b8\") " pod="openshift-marketplace/redhat-operators-j6p6n" Dec 03 16:36:01 crc kubenswrapper[4998]: I1203 16:36:01.855502 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8833f82b-b833-4a08-864f-219787f8f4b8-utilities\") pod \"redhat-operators-j6p6n\" (UID: \"8833f82b-b833-4a08-864f-219787f8f4b8\") " pod="openshift-marketplace/redhat-operators-j6p6n" Dec 03 16:36:01 crc kubenswrapper[4998]: I1203 16:36:01.957187 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8833f82b-b833-4a08-864f-219787f8f4b8-catalog-content\") pod \"redhat-operators-j6p6n\" (UID: \"8833f82b-b833-4a08-864f-219787f8f4b8\") " pod="openshift-marketplace/redhat-operators-j6p6n" Dec 03 16:36:01 crc kubenswrapper[4998]: I1203 16:36:01.957275 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cmw4\" (UniqueName: \"kubernetes.io/projected/8833f82b-b833-4a08-864f-219787f8f4b8-kube-api-access-5cmw4\") pod \"redhat-operators-j6p6n\" (UID: \"8833f82b-b833-4a08-864f-219787f8f4b8\") " pod="openshift-marketplace/redhat-operators-j6p6n" Dec 03 16:36:01 crc kubenswrapper[4998]: I1203 16:36:01.957343 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8833f82b-b833-4a08-864f-219787f8f4b8-utilities\") pod \"redhat-operators-j6p6n\" (UID: \"8833f82b-b833-4a08-864f-219787f8f4b8\") " pod="openshift-marketplace/redhat-operators-j6p6n" Dec 03 16:36:01 crc kubenswrapper[4998]: I1203 16:36:01.957679 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8833f82b-b833-4a08-864f-219787f8f4b8-catalog-content\") pod \"redhat-operators-j6p6n\" (UID: \"8833f82b-b833-4a08-864f-219787f8f4b8\") " pod="openshift-marketplace/redhat-operators-j6p6n" Dec 03 16:36:01 crc kubenswrapper[4998]: I1203 16:36:01.957917 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8833f82b-b833-4a08-864f-219787f8f4b8-utilities\") pod \"redhat-operators-j6p6n\" (UID: \"8833f82b-b833-4a08-864f-219787f8f4b8\") " pod="openshift-marketplace/redhat-operators-j6p6n" Dec 03 16:36:01 crc kubenswrapper[4998]: I1203 16:36:01.975258 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cmw4\" (UniqueName: \"kubernetes.io/projected/8833f82b-b833-4a08-864f-219787f8f4b8-kube-api-access-5cmw4\") pod \"redhat-operators-j6p6n\" (UID: \"8833f82b-b833-4a08-864f-219787f8f4b8\") " pod="openshift-marketplace/redhat-operators-j6p6n" Dec 03 16:36:02 crc kubenswrapper[4998]: I1203 16:36:02.053854 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-tdm2m"] Dec 03 16:36:02 crc kubenswrapper[4998]: I1203 16:36:02.064315 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j6p6n" Dec 03 16:36:02 crc kubenswrapper[4998]: I1203 16:36:02.066895 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-7bf7-account-create-update-xqtj9"] Dec 03 16:36:02 crc kubenswrapper[4998]: I1203 16:36:02.076101 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-tdm2m"] Dec 03 16:36:02 crc kubenswrapper[4998]: I1203 16:36:02.085518 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-7bf7-account-create-update-xqtj9"] Dec 03 16:36:02 crc kubenswrapper[4998]: I1203 16:36:02.599750 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j6p6n"] Dec 03 16:36:02 crc kubenswrapper[4998]: W1203 16:36:02.603654 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8833f82b_b833_4a08_864f_219787f8f4b8.slice/crio-184b38e2761d03832996a02e01b5a2c4770c5042f3cc28f96dd3649211c70d37 WatchSource:0}: Error finding container 184b38e2761d03832996a02e01b5a2c4770c5042f3cc28f96dd3649211c70d37: Status 404 returned error can't find the container with id 184b38e2761d03832996a02e01b5a2c4770c5042f3cc28f96dd3649211c70d37 Dec 03 16:36:03 crc kubenswrapper[4998]: I1203 16:36:03.028318 4998 generic.go:334] "Generic (PLEG): container finished" podID="742311fa-8f91-4738-ae24-66e591d9b8a1" containerID="37af099517bdd319cebc2c10d9aa950f7734fceed3ca34801eac1eec50ac3b52" exitCode=0 Dec 03 16:36:03 crc kubenswrapper[4998]: I1203 16:36:03.028388 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q" event={"ID":"742311fa-8f91-4738-ae24-66e591d9b8a1","Type":"ContainerDied","Data":"37af099517bdd319cebc2c10d9aa950f7734fceed3ca34801eac1eec50ac3b52"} Dec 03 16:36:03 crc kubenswrapper[4998]: I1203 16:36:03.030088 4998 generic.go:334] "Generic (PLEG): container finished" podID="8833f82b-b833-4a08-864f-219787f8f4b8" containerID="60af9cfe9e4c5bf377f63b6aa79ebcfd6a414a752e62c2a885cff916763cbdb1" exitCode=0 Dec 03 16:36:03 crc kubenswrapper[4998]: I1203 16:36:03.030295 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j6p6n" event={"ID":"8833f82b-b833-4a08-864f-219787f8f4b8","Type":"ContainerDied","Data":"60af9cfe9e4c5bf377f63b6aa79ebcfd6a414a752e62c2a885cff916763cbdb1"} Dec 03 16:36:03 crc kubenswrapper[4998]: I1203 16:36:03.030310 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j6p6n" event={"ID":"8833f82b-b833-4a08-864f-219787f8f4b8","Type":"ContainerStarted","Data":"184b38e2761d03832996a02e01b5a2c4770c5042f3cc28f96dd3649211c70d37"} Dec 03 16:36:03 crc kubenswrapper[4998]: I1203 16:36:03.031156 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-4825-account-create-update-k82lm"] Dec 03 16:36:03 crc kubenswrapper[4998]: I1203 16:36:03.040060 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-s9sg4"] Dec 03 16:36:03 crc kubenswrapper[4998]: I1203 16:36:03.053408 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-8rt9f"] Dec 03 16:36:03 crc kubenswrapper[4998]: I1203 16:36:03.061784 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-4825-account-create-update-k82lm"] Dec 03 16:36:03 crc kubenswrapper[4998]: I1203 16:36:03.070290 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-8cdb-account-create-update-xxcpw"] Dec 03 16:36:03 crc kubenswrapper[4998]: I1203 16:36:03.081656 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-s9sg4"] Dec 03 16:36:03 crc kubenswrapper[4998]: I1203 16:36:03.090354 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-8cdb-account-create-update-xxcpw"] Dec 03 16:36:03 crc kubenswrapper[4998]: I1203 16:36:03.098399 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-8rt9f"] Dec 03 16:36:03 crc kubenswrapper[4998]: I1203 16:36:03.700232 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e609dd6-5dee-4791-91e7-21e3191db678" path="/var/lib/kubelet/pods/0e609dd6-5dee-4791-91e7-21e3191db678/volumes" Dec 03 16:36:03 crc kubenswrapper[4998]: I1203 16:36:03.702019 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99031ef1-ff62-4058-bf84-5c2240fbce40" path="/var/lib/kubelet/pods/99031ef1-ff62-4058-bf84-5c2240fbce40/volumes" Dec 03 16:36:03 crc kubenswrapper[4998]: I1203 16:36:03.703244 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0863e3d-3fe9-4f65-8436-6e2c8cc1165c" path="/var/lib/kubelet/pods/c0863e3d-3fe9-4f65-8436-6e2c8cc1165c/volumes" Dec 03 16:36:03 crc kubenswrapper[4998]: I1203 16:36:03.704582 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c49a0bb7-935a-4d43-bd6f-848d5e44b43f" path="/var/lib/kubelet/pods/c49a0bb7-935a-4d43-bd6f-848d5e44b43f/volumes" Dec 03 16:36:03 crc kubenswrapper[4998]: I1203 16:36:03.706746 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5b99365-b407-4871-9930-1c7fa659c397" path="/var/lib/kubelet/pods/e5b99365-b407-4871-9930-1c7fa659c397/volumes" Dec 03 16:36:03 crc kubenswrapper[4998]: I1203 16:36:03.707781 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7a9cab4-be56-4a01-9b34-493370f1a56a" path="/var/lib/kubelet/pods/f7a9cab4-be56-4a01-9b34-493370f1a56a/volumes" Dec 03 16:36:04 crc kubenswrapper[4998]: I1203 16:36:04.461483 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q" Dec 03 16:36:04 crc kubenswrapper[4998]: I1203 16:36:04.607532 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6c68\" (UniqueName: \"kubernetes.io/projected/742311fa-8f91-4738-ae24-66e591d9b8a1-kube-api-access-b6c68\") pod \"742311fa-8f91-4738-ae24-66e591d9b8a1\" (UID: \"742311fa-8f91-4738-ae24-66e591d9b8a1\") " Dec 03 16:36:04 crc kubenswrapper[4998]: I1203 16:36:04.607629 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/742311fa-8f91-4738-ae24-66e591d9b8a1-ssh-key\") pod \"742311fa-8f91-4738-ae24-66e591d9b8a1\" (UID: \"742311fa-8f91-4738-ae24-66e591d9b8a1\") " Dec 03 16:36:04 crc kubenswrapper[4998]: I1203 16:36:04.607791 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/742311fa-8f91-4738-ae24-66e591d9b8a1-inventory\") pod \"742311fa-8f91-4738-ae24-66e591d9b8a1\" (UID: \"742311fa-8f91-4738-ae24-66e591d9b8a1\") " Dec 03 16:36:04 crc kubenswrapper[4998]: I1203 16:36:04.619013 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/742311fa-8f91-4738-ae24-66e591d9b8a1-kube-api-access-b6c68" (OuterVolumeSpecName: "kube-api-access-b6c68") pod "742311fa-8f91-4738-ae24-66e591d9b8a1" (UID: "742311fa-8f91-4738-ae24-66e591d9b8a1"). InnerVolumeSpecName "kube-api-access-b6c68". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:36:04 crc kubenswrapper[4998]: I1203 16:36:04.642992 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/742311fa-8f91-4738-ae24-66e591d9b8a1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "742311fa-8f91-4738-ae24-66e591d9b8a1" (UID: "742311fa-8f91-4738-ae24-66e591d9b8a1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:36:04 crc kubenswrapper[4998]: I1203 16:36:04.648681 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/742311fa-8f91-4738-ae24-66e591d9b8a1-inventory" (OuterVolumeSpecName: "inventory") pod "742311fa-8f91-4738-ae24-66e591d9b8a1" (UID: "742311fa-8f91-4738-ae24-66e591d9b8a1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:36:04 crc kubenswrapper[4998]: I1203 16:36:04.710645 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6c68\" (UniqueName: \"kubernetes.io/projected/742311fa-8f91-4738-ae24-66e591d9b8a1-kube-api-access-b6c68\") on node \"crc\" DevicePath \"\"" Dec 03 16:36:04 crc kubenswrapper[4998]: I1203 16:36:04.710883 4998 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/742311fa-8f91-4738-ae24-66e591d9b8a1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 16:36:04 crc kubenswrapper[4998]: I1203 16:36:04.710977 4998 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/742311fa-8f91-4738-ae24-66e591d9b8a1-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 16:36:05 crc kubenswrapper[4998]: I1203 16:36:05.054250 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j6p6n" event={"ID":"8833f82b-b833-4a08-864f-219787f8f4b8","Type":"ContainerStarted","Data":"b245804c1c56fd915fcb5e9da655a07fe1a80be1c3cd2283180661e439068bab"} Dec 03 16:36:05 crc kubenswrapper[4998]: I1203 16:36:05.056037 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q" event={"ID":"742311fa-8f91-4738-ae24-66e591d9b8a1","Type":"ContainerDied","Data":"6e29f7e69a77627695fc2394d8dbe23415068406929314ce5789560bd072fdf0"} Dec 03 16:36:05 crc kubenswrapper[4998]: I1203 16:36:05.056102 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e29f7e69a77627695fc2394d8dbe23415068406929314ce5789560bd072fdf0" Dec 03 16:36:05 crc kubenswrapper[4998]: I1203 16:36:05.056067 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q" Dec 03 16:36:05 crc kubenswrapper[4998]: I1203 16:36:05.126730 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk"] Dec 03 16:36:05 crc kubenswrapper[4998]: E1203 16:36:05.127287 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="742311fa-8f91-4738-ae24-66e591d9b8a1" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 03 16:36:05 crc kubenswrapper[4998]: I1203 16:36:05.127308 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="742311fa-8f91-4738-ae24-66e591d9b8a1" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 03 16:36:05 crc kubenswrapper[4998]: I1203 16:36:05.127582 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="742311fa-8f91-4738-ae24-66e591d9b8a1" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 03 16:36:05 crc kubenswrapper[4998]: I1203 16:36:05.128486 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk" Dec 03 16:36:05 crc kubenswrapper[4998]: I1203 16:36:05.130685 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 16:36:05 crc kubenswrapper[4998]: I1203 16:36:05.131083 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 16:36:05 crc kubenswrapper[4998]: I1203 16:36:05.131571 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 16:36:05 crc kubenswrapper[4998]: I1203 16:36:05.136218 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kcj9x" Dec 03 16:36:05 crc kubenswrapper[4998]: I1203 16:36:05.140731 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk"] Dec 03 16:36:05 crc kubenswrapper[4998]: I1203 16:36:05.226156 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fb49d67-f5a0-4728-b3e1-dc596955f03a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk\" (UID: \"5fb49d67-f5a0-4728-b3e1-dc596955f03a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk" Dec 03 16:36:05 crc kubenswrapper[4998]: I1203 16:36:05.226358 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6hbb\" (UniqueName: \"kubernetes.io/projected/5fb49d67-f5a0-4728-b3e1-dc596955f03a-kube-api-access-d6hbb\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk\" (UID: \"5fb49d67-f5a0-4728-b3e1-dc596955f03a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk" Dec 03 16:36:05 crc kubenswrapper[4998]: I1203 16:36:05.226406 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fb49d67-f5a0-4728-b3e1-dc596955f03a-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk\" (UID: \"5fb49d67-f5a0-4728-b3e1-dc596955f03a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk" Dec 03 16:36:05 crc kubenswrapper[4998]: I1203 16:36:05.328543 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6hbb\" (UniqueName: \"kubernetes.io/projected/5fb49d67-f5a0-4728-b3e1-dc596955f03a-kube-api-access-d6hbb\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk\" (UID: \"5fb49d67-f5a0-4728-b3e1-dc596955f03a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk" Dec 03 16:36:05 crc kubenswrapper[4998]: I1203 16:36:05.328609 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fb49d67-f5a0-4728-b3e1-dc596955f03a-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk\" (UID: \"5fb49d67-f5a0-4728-b3e1-dc596955f03a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk" Dec 03 16:36:05 crc kubenswrapper[4998]: I1203 16:36:05.328668 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fb49d67-f5a0-4728-b3e1-dc596955f03a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk\" (UID: \"5fb49d67-f5a0-4728-b3e1-dc596955f03a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk" Dec 03 16:36:05 crc kubenswrapper[4998]: I1203 16:36:05.333441 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fb49d67-f5a0-4728-b3e1-dc596955f03a-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk\" (UID: \"5fb49d67-f5a0-4728-b3e1-dc596955f03a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk" Dec 03 16:36:05 crc kubenswrapper[4998]: I1203 16:36:05.348436 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fb49d67-f5a0-4728-b3e1-dc596955f03a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk\" (UID: \"5fb49d67-f5a0-4728-b3e1-dc596955f03a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk" Dec 03 16:36:05 crc kubenswrapper[4998]: I1203 16:36:05.354504 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6hbb\" (UniqueName: \"kubernetes.io/projected/5fb49d67-f5a0-4728-b3e1-dc596955f03a-kube-api-access-d6hbb\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk\" (UID: \"5fb49d67-f5a0-4728-b3e1-dc596955f03a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk" Dec 03 16:36:05 crc kubenswrapper[4998]: I1203 16:36:05.447897 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk" Dec 03 16:36:05 crc kubenswrapper[4998]: I1203 16:36:05.985228 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk"] Dec 03 16:36:05 crc kubenswrapper[4998]: W1203 16:36:05.985606 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5fb49d67_f5a0_4728_b3e1_dc596955f03a.slice/crio-ef42397c9adb675585fa0c3688e69d1ef06b866ab553220984abadbf6a274fe2 WatchSource:0}: Error finding container ef42397c9adb675585fa0c3688e69d1ef06b866ab553220984abadbf6a274fe2: Status 404 returned error can't find the container with id ef42397c9adb675585fa0c3688e69d1ef06b866ab553220984abadbf6a274fe2 Dec 03 16:36:06 crc kubenswrapper[4998]: I1203 16:36:06.065489 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk" event={"ID":"5fb49d67-f5a0-4728-b3e1-dc596955f03a","Type":"ContainerStarted","Data":"ef42397c9adb675585fa0c3688e69d1ef06b866ab553220984abadbf6a274fe2"} Dec 03 16:36:07 crc kubenswrapper[4998]: I1203 16:36:07.075935 4998 generic.go:334] "Generic (PLEG): container finished" podID="8833f82b-b833-4a08-864f-219787f8f4b8" containerID="b245804c1c56fd915fcb5e9da655a07fe1a80be1c3cd2283180661e439068bab" exitCode=0 Dec 03 16:36:07 crc kubenswrapper[4998]: I1203 16:36:07.076019 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j6p6n" event={"ID":"8833f82b-b833-4a08-864f-219787f8f4b8","Type":"ContainerDied","Data":"b245804c1c56fd915fcb5e9da655a07fe1a80be1c3cd2283180661e439068bab"} Dec 03 16:36:08 crc kubenswrapper[4998]: I1203 16:36:08.086897 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk" event={"ID":"5fb49d67-f5a0-4728-b3e1-dc596955f03a","Type":"ContainerStarted","Data":"b508b0a17e7e213711f36fc46f1c3362a31f3a6bbc3583bc9e9d578e75ff0094"} Dec 03 16:36:08 crc kubenswrapper[4998]: I1203 16:36:08.090323 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j6p6n" event={"ID":"8833f82b-b833-4a08-864f-219787f8f4b8","Type":"ContainerStarted","Data":"ce1c59c739406cdb93cad25c7bb132c945faec44740d4be70e76421abd5a69e8"} Dec 03 16:36:08 crc kubenswrapper[4998]: I1203 16:36:08.115441 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk" podStartSLOduration=1.746154879 podStartE2EDuration="3.115424118s" podCreationTimestamp="2025-12-03 16:36:05 +0000 UTC" firstStartedPulling="2025-12-03 16:36:05.989085634 +0000 UTC m=+1944.600785857" lastFinishedPulling="2025-12-03 16:36:07.358354873 +0000 UTC m=+1945.970055096" observedRunningTime="2025-12-03 16:36:08.109551121 +0000 UTC m=+1946.721251354" watchObservedRunningTime="2025-12-03 16:36:08.115424118 +0000 UTC m=+1946.727124351" Dec 03 16:36:08 crc kubenswrapper[4998]: I1203 16:36:08.134689 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j6p6n" podStartSLOduration=2.56452775 podStartE2EDuration="7.134673977s" podCreationTimestamp="2025-12-03 16:36:01 +0000 UTC" firstStartedPulling="2025-12-03 16:36:03.032208896 +0000 UTC m=+1941.643909119" lastFinishedPulling="2025-12-03 16:36:07.602355073 +0000 UTC m=+1946.214055346" observedRunningTime="2025-12-03 16:36:08.130138704 +0000 UTC m=+1946.741838927" watchObservedRunningTime="2025-12-03 16:36:08.134673977 +0000 UTC m=+1946.746374200" Dec 03 16:36:10 crc kubenswrapper[4998]: I1203 16:36:10.457514 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-667sw" Dec 03 16:36:12 crc kubenswrapper[4998]: I1203 16:36:12.065200 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j6p6n" Dec 03 16:36:12 crc kubenswrapper[4998]: I1203 16:36:12.065468 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j6p6n" Dec 03 16:36:12 crc kubenswrapper[4998]: I1203 16:36:12.240197 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-667sw"] Dec 03 16:36:12 crc kubenswrapper[4998]: I1203 16:36:12.240485 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-667sw" podUID="fc03c421-4ea1-46c9-ae67-03af108bf717" containerName="registry-server" containerID="cri-o://7c35ec5d9e0f8d3c60f967ceac7eb7e26613ef6c4899ca9f7f501556135571b6" gracePeriod=2 Dec 03 16:36:12 crc kubenswrapper[4998]: I1203 16:36:12.718543 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-667sw" Dec 03 16:36:12 crc kubenswrapper[4998]: I1203 16:36:12.875887 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc03c421-4ea1-46c9-ae67-03af108bf717-catalog-content\") pod \"fc03c421-4ea1-46c9-ae67-03af108bf717\" (UID: \"fc03c421-4ea1-46c9-ae67-03af108bf717\") " Dec 03 16:36:12 crc kubenswrapper[4998]: I1203 16:36:12.876633 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc03c421-4ea1-46c9-ae67-03af108bf717-utilities\") pod \"fc03c421-4ea1-46c9-ae67-03af108bf717\" (UID: \"fc03c421-4ea1-46c9-ae67-03af108bf717\") " Dec 03 16:36:12 crc kubenswrapper[4998]: I1203 16:36:12.876840 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6lzd\" (UniqueName: \"kubernetes.io/projected/fc03c421-4ea1-46c9-ae67-03af108bf717-kube-api-access-d6lzd\") pod \"fc03c421-4ea1-46c9-ae67-03af108bf717\" (UID: \"fc03c421-4ea1-46c9-ae67-03af108bf717\") " Dec 03 16:36:12 crc kubenswrapper[4998]: I1203 16:36:12.877141 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc03c421-4ea1-46c9-ae67-03af108bf717-utilities" (OuterVolumeSpecName: "utilities") pod "fc03c421-4ea1-46c9-ae67-03af108bf717" (UID: "fc03c421-4ea1-46c9-ae67-03af108bf717"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:36:12 crc kubenswrapper[4998]: I1203 16:36:12.877563 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc03c421-4ea1-46c9-ae67-03af108bf717-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:36:12 crc kubenswrapper[4998]: I1203 16:36:12.882112 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc03c421-4ea1-46c9-ae67-03af108bf717-kube-api-access-d6lzd" (OuterVolumeSpecName: "kube-api-access-d6lzd") pod "fc03c421-4ea1-46c9-ae67-03af108bf717" (UID: "fc03c421-4ea1-46c9-ae67-03af108bf717"). InnerVolumeSpecName "kube-api-access-d6lzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:36:12 crc kubenswrapper[4998]: I1203 16:36:12.898095 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc03c421-4ea1-46c9-ae67-03af108bf717-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fc03c421-4ea1-46c9-ae67-03af108bf717" (UID: "fc03c421-4ea1-46c9-ae67-03af108bf717"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:36:12 crc kubenswrapper[4998]: I1203 16:36:12.980198 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc03c421-4ea1-46c9-ae67-03af108bf717-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:36:12 crc kubenswrapper[4998]: I1203 16:36:12.980555 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6lzd\" (UniqueName: \"kubernetes.io/projected/fc03c421-4ea1-46c9-ae67-03af108bf717-kube-api-access-d6lzd\") on node \"crc\" DevicePath \"\"" Dec 03 16:36:13 crc kubenswrapper[4998]: I1203 16:36:13.115397 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j6p6n" podUID="8833f82b-b833-4a08-864f-219787f8f4b8" containerName="registry-server" probeResult="failure" output=< Dec 03 16:36:13 crc kubenswrapper[4998]: timeout: failed to connect service ":50051" within 1s Dec 03 16:36:13 crc kubenswrapper[4998]: > Dec 03 16:36:13 crc kubenswrapper[4998]: I1203 16:36:13.136118 4998 generic.go:334] "Generic (PLEG): container finished" podID="fc03c421-4ea1-46c9-ae67-03af108bf717" containerID="7c35ec5d9e0f8d3c60f967ceac7eb7e26613ef6c4899ca9f7f501556135571b6" exitCode=0 Dec 03 16:36:13 crc kubenswrapper[4998]: I1203 16:36:13.136162 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-667sw" event={"ID":"fc03c421-4ea1-46c9-ae67-03af108bf717","Type":"ContainerDied","Data":"7c35ec5d9e0f8d3c60f967ceac7eb7e26613ef6c4899ca9f7f501556135571b6"} Dec 03 16:36:13 crc kubenswrapper[4998]: I1203 16:36:13.136179 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-667sw" Dec 03 16:36:13 crc kubenswrapper[4998]: I1203 16:36:13.136192 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-667sw" event={"ID":"fc03c421-4ea1-46c9-ae67-03af108bf717","Type":"ContainerDied","Data":"af6719fbb0277d3806d982b2b721909c6a5d79fb0b270c178f2347de91651ae2"} Dec 03 16:36:13 crc kubenswrapper[4998]: I1203 16:36:13.136212 4998 scope.go:117] "RemoveContainer" containerID="7c35ec5d9e0f8d3c60f967ceac7eb7e26613ef6c4899ca9f7f501556135571b6" Dec 03 16:36:13 crc kubenswrapper[4998]: I1203 16:36:13.155271 4998 scope.go:117] "RemoveContainer" containerID="eb781609f60403f61d92e854e2c3566c29913d5b2205b97365077cc9d21bca07" Dec 03 16:36:13 crc kubenswrapper[4998]: I1203 16:36:13.168703 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-667sw"] Dec 03 16:36:13 crc kubenswrapper[4998]: I1203 16:36:13.181321 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-667sw"] Dec 03 16:36:13 crc kubenswrapper[4998]: I1203 16:36:13.204552 4998 scope.go:117] "RemoveContainer" containerID="010d3d4ab8f88b11b226c20f8044c10b3cd9cce526bce6084871fddd298b52c4" Dec 03 16:36:13 crc kubenswrapper[4998]: I1203 16:36:13.235020 4998 scope.go:117] "RemoveContainer" containerID="7c35ec5d9e0f8d3c60f967ceac7eb7e26613ef6c4899ca9f7f501556135571b6" Dec 03 16:36:13 crc kubenswrapper[4998]: E1203 16:36:13.235510 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c35ec5d9e0f8d3c60f967ceac7eb7e26613ef6c4899ca9f7f501556135571b6\": container with ID starting with 7c35ec5d9e0f8d3c60f967ceac7eb7e26613ef6c4899ca9f7f501556135571b6 not found: ID does not exist" containerID="7c35ec5d9e0f8d3c60f967ceac7eb7e26613ef6c4899ca9f7f501556135571b6" Dec 03 16:36:13 crc kubenswrapper[4998]: I1203 16:36:13.235632 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c35ec5d9e0f8d3c60f967ceac7eb7e26613ef6c4899ca9f7f501556135571b6"} err="failed to get container status \"7c35ec5d9e0f8d3c60f967ceac7eb7e26613ef6c4899ca9f7f501556135571b6\": rpc error: code = NotFound desc = could not find container \"7c35ec5d9e0f8d3c60f967ceac7eb7e26613ef6c4899ca9f7f501556135571b6\": container with ID starting with 7c35ec5d9e0f8d3c60f967ceac7eb7e26613ef6c4899ca9f7f501556135571b6 not found: ID does not exist" Dec 03 16:36:13 crc kubenswrapper[4998]: I1203 16:36:13.235746 4998 scope.go:117] "RemoveContainer" containerID="eb781609f60403f61d92e854e2c3566c29913d5b2205b97365077cc9d21bca07" Dec 03 16:36:13 crc kubenswrapper[4998]: E1203 16:36:13.236163 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb781609f60403f61d92e854e2c3566c29913d5b2205b97365077cc9d21bca07\": container with ID starting with eb781609f60403f61d92e854e2c3566c29913d5b2205b97365077cc9d21bca07 not found: ID does not exist" containerID="eb781609f60403f61d92e854e2c3566c29913d5b2205b97365077cc9d21bca07" Dec 03 16:36:13 crc kubenswrapper[4998]: I1203 16:36:13.236215 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb781609f60403f61d92e854e2c3566c29913d5b2205b97365077cc9d21bca07"} err="failed to get container status \"eb781609f60403f61d92e854e2c3566c29913d5b2205b97365077cc9d21bca07\": rpc error: code = NotFound desc = could not find container \"eb781609f60403f61d92e854e2c3566c29913d5b2205b97365077cc9d21bca07\": container with ID starting with eb781609f60403f61d92e854e2c3566c29913d5b2205b97365077cc9d21bca07 not found: ID does not exist" Dec 03 16:36:13 crc kubenswrapper[4998]: I1203 16:36:13.236245 4998 scope.go:117] "RemoveContainer" containerID="010d3d4ab8f88b11b226c20f8044c10b3cd9cce526bce6084871fddd298b52c4" Dec 03 16:36:13 crc kubenswrapper[4998]: E1203 16:36:13.236574 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"010d3d4ab8f88b11b226c20f8044c10b3cd9cce526bce6084871fddd298b52c4\": container with ID starting with 010d3d4ab8f88b11b226c20f8044c10b3cd9cce526bce6084871fddd298b52c4 not found: ID does not exist" containerID="010d3d4ab8f88b11b226c20f8044c10b3cd9cce526bce6084871fddd298b52c4" Dec 03 16:36:13 crc kubenswrapper[4998]: I1203 16:36:13.236907 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"010d3d4ab8f88b11b226c20f8044c10b3cd9cce526bce6084871fddd298b52c4"} err="failed to get container status \"010d3d4ab8f88b11b226c20f8044c10b3cd9cce526bce6084871fddd298b52c4\": rpc error: code = NotFound desc = could not find container \"010d3d4ab8f88b11b226c20f8044c10b3cd9cce526bce6084871fddd298b52c4\": container with ID starting with 010d3d4ab8f88b11b226c20f8044c10b3cd9cce526bce6084871fddd298b52c4 not found: ID does not exist" Dec 03 16:36:13 crc kubenswrapper[4998]: I1203 16:36:13.691212 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc03c421-4ea1-46c9-ae67-03af108bf717" path="/var/lib/kubelet/pods/fc03c421-4ea1-46c9-ae67-03af108bf717/volumes" Dec 03 16:36:22 crc kubenswrapper[4998]: I1203 16:36:22.138639 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j6p6n" Dec 03 16:36:22 crc kubenswrapper[4998]: I1203 16:36:22.201481 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j6p6n" Dec 03 16:36:22 crc kubenswrapper[4998]: I1203 16:36:22.374867 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j6p6n"] Dec 03 16:36:23 crc kubenswrapper[4998]: I1203 16:36:23.227149 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-j6p6n" podUID="8833f82b-b833-4a08-864f-219787f8f4b8" containerName="registry-server" containerID="cri-o://ce1c59c739406cdb93cad25c7bb132c945faec44740d4be70e76421abd5a69e8" gracePeriod=2 Dec 03 16:36:23 crc kubenswrapper[4998]: I1203 16:36:23.662136 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j6p6n" Dec 03 16:36:23 crc kubenswrapper[4998]: I1203 16:36:23.822285 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8833f82b-b833-4a08-864f-219787f8f4b8-utilities\") pod \"8833f82b-b833-4a08-864f-219787f8f4b8\" (UID: \"8833f82b-b833-4a08-864f-219787f8f4b8\") " Dec 03 16:36:23 crc kubenswrapper[4998]: I1203 16:36:23.822442 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8833f82b-b833-4a08-864f-219787f8f4b8-catalog-content\") pod \"8833f82b-b833-4a08-864f-219787f8f4b8\" (UID: \"8833f82b-b833-4a08-864f-219787f8f4b8\") " Dec 03 16:36:23 crc kubenswrapper[4998]: I1203 16:36:23.823069 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8833f82b-b833-4a08-864f-219787f8f4b8-utilities" (OuterVolumeSpecName: "utilities") pod "8833f82b-b833-4a08-864f-219787f8f4b8" (UID: "8833f82b-b833-4a08-864f-219787f8f4b8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:36:23 crc kubenswrapper[4998]: I1203 16:36:23.826049 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5cmw4\" (UniqueName: \"kubernetes.io/projected/8833f82b-b833-4a08-864f-219787f8f4b8-kube-api-access-5cmw4\") pod \"8833f82b-b833-4a08-864f-219787f8f4b8\" (UID: \"8833f82b-b833-4a08-864f-219787f8f4b8\") " Dec 03 16:36:23 crc kubenswrapper[4998]: I1203 16:36:23.826921 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8833f82b-b833-4a08-864f-219787f8f4b8-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:36:23 crc kubenswrapper[4998]: I1203 16:36:23.836574 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8833f82b-b833-4a08-864f-219787f8f4b8-kube-api-access-5cmw4" (OuterVolumeSpecName: "kube-api-access-5cmw4") pod "8833f82b-b833-4a08-864f-219787f8f4b8" (UID: "8833f82b-b833-4a08-864f-219787f8f4b8"). InnerVolumeSpecName "kube-api-access-5cmw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:36:23 crc kubenswrapper[4998]: I1203 16:36:23.926036 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8833f82b-b833-4a08-864f-219787f8f4b8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8833f82b-b833-4a08-864f-219787f8f4b8" (UID: "8833f82b-b833-4a08-864f-219787f8f4b8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:36:23 crc kubenswrapper[4998]: I1203 16:36:23.929274 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5cmw4\" (UniqueName: \"kubernetes.io/projected/8833f82b-b833-4a08-864f-219787f8f4b8-kube-api-access-5cmw4\") on node \"crc\" DevicePath \"\"" Dec 03 16:36:23 crc kubenswrapper[4998]: I1203 16:36:23.929319 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8833f82b-b833-4a08-864f-219787f8f4b8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:36:24 crc kubenswrapper[4998]: I1203 16:36:24.240865 4998 generic.go:334] "Generic (PLEG): container finished" podID="8833f82b-b833-4a08-864f-219787f8f4b8" containerID="ce1c59c739406cdb93cad25c7bb132c945faec44740d4be70e76421abd5a69e8" exitCode=0 Dec 03 16:36:24 crc kubenswrapper[4998]: I1203 16:36:24.240911 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j6p6n" event={"ID":"8833f82b-b833-4a08-864f-219787f8f4b8","Type":"ContainerDied","Data":"ce1c59c739406cdb93cad25c7bb132c945faec44740d4be70e76421abd5a69e8"} Dec 03 16:36:24 crc kubenswrapper[4998]: I1203 16:36:24.240941 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j6p6n" event={"ID":"8833f82b-b833-4a08-864f-219787f8f4b8","Type":"ContainerDied","Data":"184b38e2761d03832996a02e01b5a2c4770c5042f3cc28f96dd3649211c70d37"} Dec 03 16:36:24 crc kubenswrapper[4998]: I1203 16:36:24.240959 4998 scope.go:117] "RemoveContainer" containerID="ce1c59c739406cdb93cad25c7bb132c945faec44740d4be70e76421abd5a69e8" Dec 03 16:36:24 crc kubenswrapper[4998]: I1203 16:36:24.241091 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j6p6n" Dec 03 16:36:24 crc kubenswrapper[4998]: I1203 16:36:24.273773 4998 scope.go:117] "RemoveContainer" containerID="b245804c1c56fd915fcb5e9da655a07fe1a80be1c3cd2283180661e439068bab" Dec 03 16:36:24 crc kubenswrapper[4998]: I1203 16:36:24.282551 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j6p6n"] Dec 03 16:36:24 crc kubenswrapper[4998]: I1203 16:36:24.294862 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-j6p6n"] Dec 03 16:36:24 crc kubenswrapper[4998]: I1203 16:36:24.308818 4998 scope.go:117] "RemoveContainer" containerID="60af9cfe9e4c5bf377f63b6aa79ebcfd6a414a752e62c2a885cff916763cbdb1" Dec 03 16:36:24 crc kubenswrapper[4998]: I1203 16:36:24.365192 4998 scope.go:117] "RemoveContainer" containerID="ce1c59c739406cdb93cad25c7bb132c945faec44740d4be70e76421abd5a69e8" Dec 03 16:36:24 crc kubenswrapper[4998]: E1203 16:36:24.366435 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce1c59c739406cdb93cad25c7bb132c945faec44740d4be70e76421abd5a69e8\": container with ID starting with ce1c59c739406cdb93cad25c7bb132c945faec44740d4be70e76421abd5a69e8 not found: ID does not exist" containerID="ce1c59c739406cdb93cad25c7bb132c945faec44740d4be70e76421abd5a69e8" Dec 03 16:36:24 crc kubenswrapper[4998]: I1203 16:36:24.366470 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce1c59c739406cdb93cad25c7bb132c945faec44740d4be70e76421abd5a69e8"} err="failed to get container status \"ce1c59c739406cdb93cad25c7bb132c945faec44740d4be70e76421abd5a69e8\": rpc error: code = NotFound desc = could not find container \"ce1c59c739406cdb93cad25c7bb132c945faec44740d4be70e76421abd5a69e8\": container with ID starting with ce1c59c739406cdb93cad25c7bb132c945faec44740d4be70e76421abd5a69e8 not found: ID does not exist" Dec 03 16:36:24 crc kubenswrapper[4998]: I1203 16:36:24.366507 4998 scope.go:117] "RemoveContainer" containerID="b245804c1c56fd915fcb5e9da655a07fe1a80be1c3cd2283180661e439068bab" Dec 03 16:36:24 crc kubenswrapper[4998]: E1203 16:36:24.367979 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b245804c1c56fd915fcb5e9da655a07fe1a80be1c3cd2283180661e439068bab\": container with ID starting with b245804c1c56fd915fcb5e9da655a07fe1a80be1c3cd2283180661e439068bab not found: ID does not exist" containerID="b245804c1c56fd915fcb5e9da655a07fe1a80be1c3cd2283180661e439068bab" Dec 03 16:36:24 crc kubenswrapper[4998]: I1203 16:36:24.368003 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b245804c1c56fd915fcb5e9da655a07fe1a80be1c3cd2283180661e439068bab"} err="failed to get container status \"b245804c1c56fd915fcb5e9da655a07fe1a80be1c3cd2283180661e439068bab\": rpc error: code = NotFound desc = could not find container \"b245804c1c56fd915fcb5e9da655a07fe1a80be1c3cd2283180661e439068bab\": container with ID starting with b245804c1c56fd915fcb5e9da655a07fe1a80be1c3cd2283180661e439068bab not found: ID does not exist" Dec 03 16:36:24 crc kubenswrapper[4998]: I1203 16:36:24.368016 4998 scope.go:117] "RemoveContainer" containerID="60af9cfe9e4c5bf377f63b6aa79ebcfd6a414a752e62c2a885cff916763cbdb1" Dec 03 16:36:24 crc kubenswrapper[4998]: E1203 16:36:24.368398 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60af9cfe9e4c5bf377f63b6aa79ebcfd6a414a752e62c2a885cff916763cbdb1\": container with ID starting with 60af9cfe9e4c5bf377f63b6aa79ebcfd6a414a752e62c2a885cff916763cbdb1 not found: ID does not exist" containerID="60af9cfe9e4c5bf377f63b6aa79ebcfd6a414a752e62c2a885cff916763cbdb1" Dec 03 16:36:24 crc kubenswrapper[4998]: I1203 16:36:24.368428 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60af9cfe9e4c5bf377f63b6aa79ebcfd6a414a752e62c2a885cff916763cbdb1"} err="failed to get container status \"60af9cfe9e4c5bf377f63b6aa79ebcfd6a414a752e62c2a885cff916763cbdb1\": rpc error: code = NotFound desc = could not find container \"60af9cfe9e4c5bf377f63b6aa79ebcfd6a414a752e62c2a885cff916763cbdb1\": container with ID starting with 60af9cfe9e4c5bf377f63b6aa79ebcfd6a414a752e62c2a885cff916763cbdb1 not found: ID does not exist" Dec 03 16:36:25 crc kubenswrapper[4998]: I1203 16:36:25.717273 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8833f82b-b833-4a08-864f-219787f8f4b8" path="/var/lib/kubelet/pods/8833f82b-b833-4a08-864f-219787f8f4b8/volumes" Dec 03 16:36:27 crc kubenswrapper[4998]: I1203 16:36:27.111347 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:36:27 crc kubenswrapper[4998]: I1203 16:36:27.111748 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:36:48 crc kubenswrapper[4998]: I1203 16:36:48.067537 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6zkx5"] Dec 03 16:36:48 crc kubenswrapper[4998]: I1203 16:36:48.090114 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-6zkx5"] Dec 03 16:36:49 crc kubenswrapper[4998]: I1203 16:36:49.691864 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ff4ba4e-9b43-41fd-902b-977b9cc3aba9" path="/var/lib/kubelet/pods/3ff4ba4e-9b43-41fd-902b-977b9cc3aba9/volumes" Dec 03 16:36:57 crc kubenswrapper[4998]: I1203 16:36:57.111136 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:36:57 crc kubenswrapper[4998]: I1203 16:36:57.111464 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:36:57 crc kubenswrapper[4998]: I1203 16:36:57.111508 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:36:57 crc kubenswrapper[4998]: I1203 16:36:57.112212 4998 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9001df2bc0dc33071d80330946bfdea9e78811028533587a680aea91a458fee6"} pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 16:36:57 crc kubenswrapper[4998]: I1203 16:36:57.112267 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" containerID="cri-o://9001df2bc0dc33071d80330946bfdea9e78811028533587a680aea91a458fee6" gracePeriod=600 Dec 03 16:36:57 crc kubenswrapper[4998]: I1203 16:36:57.613271 4998 generic.go:334] "Generic (PLEG): container finished" podID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerID="9001df2bc0dc33071d80330946bfdea9e78811028533587a680aea91a458fee6" exitCode=0 Dec 03 16:36:57 crc kubenswrapper[4998]: I1203 16:36:57.613329 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerDied","Data":"9001df2bc0dc33071d80330946bfdea9e78811028533587a680aea91a458fee6"} Dec 03 16:36:57 crc kubenswrapper[4998]: I1203 16:36:57.613804 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155"} Dec 03 16:36:57 crc kubenswrapper[4998]: I1203 16:36:57.613853 4998 scope.go:117] "RemoveContainer" containerID="d0825218d71008ed59851edf815f54ff889bbb4bc432d9f0b6c3212511e62379" Dec 03 16:36:59 crc kubenswrapper[4998]: I1203 16:36:59.810677 4998 scope.go:117] "RemoveContainer" containerID="65a19393bd94e0171b263597878f2a7b35cfe646f0319e3410af84795657581a" Dec 03 16:36:59 crc kubenswrapper[4998]: I1203 16:36:59.853693 4998 scope.go:117] "RemoveContainer" containerID="ef79e83bb317586b636615f53ef93881cfbe790b6f3bb1a7c4e2d0776b1b189a" Dec 03 16:36:59 crc kubenswrapper[4998]: I1203 16:36:59.907710 4998 scope.go:117] "RemoveContainer" containerID="523211a913430c4a44b7d89513310d008be76421fe3f439806c255ab5d8d8db6" Dec 03 16:36:59 crc kubenswrapper[4998]: I1203 16:36:59.971720 4998 scope.go:117] "RemoveContainer" containerID="aaf5b11bcc4c4e392a0d553686315edf17fa6e88837b90d37cec482da29159f7" Dec 03 16:37:00 crc kubenswrapper[4998]: I1203 16:37:00.022042 4998 scope.go:117] "RemoveContainer" containerID="2c4799dce3f2dcce67b1ade1fac3354b0087f5bc5469a7db466cc3741baad1d4" Dec 03 16:37:00 crc kubenswrapper[4998]: I1203 16:37:00.063116 4998 scope.go:117] "RemoveContainer" containerID="c4fe18fac933af612bc197d41649650f29ab6e8496268bd2342d1c84935c7234" Dec 03 16:37:00 crc kubenswrapper[4998]: I1203 16:37:00.134656 4998 scope.go:117] "RemoveContainer" containerID="ce3e22e7f15f44efe3fb113b2f017f3130bb916c182926c239ef9495a5b26451" Dec 03 16:37:12 crc kubenswrapper[4998]: I1203 16:37:12.038639 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-jzsvl"] Dec 03 16:37:12 crc kubenswrapper[4998]: I1203 16:37:12.047958 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-jzsvl"] Dec 03 16:37:13 crc kubenswrapper[4998]: I1203 16:37:13.692293 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d" path="/var/lib/kubelet/pods/0e4d81d9-be5b-4d11-97d1-4aaa2c2d386d/volumes" Dec 03 16:37:19 crc kubenswrapper[4998]: I1203 16:37:19.061032 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vvjng"] Dec 03 16:37:19 crc kubenswrapper[4998]: I1203 16:37:19.074722 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vvjng"] Dec 03 16:37:19 crc kubenswrapper[4998]: I1203 16:37:19.697460 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86d54b38-8913-4d18-92ee-490aceef1677" path="/var/lib/kubelet/pods/86d54b38-8913-4d18-92ee-490aceef1677/volumes" Dec 03 16:37:51 crc kubenswrapper[4998]: I1203 16:37:51.103622 4998 generic.go:334] "Generic (PLEG): container finished" podID="5fb49d67-f5a0-4728-b3e1-dc596955f03a" containerID="b508b0a17e7e213711f36fc46f1c3362a31f3a6bbc3583bc9e9d578e75ff0094" exitCode=0 Dec 03 16:37:51 crc kubenswrapper[4998]: I1203 16:37:51.103679 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk" event={"ID":"5fb49d67-f5a0-4728-b3e1-dc596955f03a","Type":"ContainerDied","Data":"b508b0a17e7e213711f36fc46f1c3362a31f3a6bbc3583bc9e9d578e75ff0094"} Dec 03 16:37:52 crc kubenswrapper[4998]: I1203 16:37:52.537828 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk" Dec 03 16:37:52 crc kubenswrapper[4998]: I1203 16:37:52.659670 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fb49d67-f5a0-4728-b3e1-dc596955f03a-inventory\") pod \"5fb49d67-f5a0-4728-b3e1-dc596955f03a\" (UID: \"5fb49d67-f5a0-4728-b3e1-dc596955f03a\") " Dec 03 16:37:52 crc kubenswrapper[4998]: I1203 16:37:52.660159 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fb49d67-f5a0-4728-b3e1-dc596955f03a-ssh-key\") pod \"5fb49d67-f5a0-4728-b3e1-dc596955f03a\" (UID: \"5fb49d67-f5a0-4728-b3e1-dc596955f03a\") " Dec 03 16:37:52 crc kubenswrapper[4998]: I1203 16:37:52.660197 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6hbb\" (UniqueName: \"kubernetes.io/projected/5fb49d67-f5a0-4728-b3e1-dc596955f03a-kube-api-access-d6hbb\") pod \"5fb49d67-f5a0-4728-b3e1-dc596955f03a\" (UID: \"5fb49d67-f5a0-4728-b3e1-dc596955f03a\") " Dec 03 16:37:52 crc kubenswrapper[4998]: I1203 16:37:52.672120 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fb49d67-f5a0-4728-b3e1-dc596955f03a-kube-api-access-d6hbb" (OuterVolumeSpecName: "kube-api-access-d6hbb") pod "5fb49d67-f5a0-4728-b3e1-dc596955f03a" (UID: "5fb49d67-f5a0-4728-b3e1-dc596955f03a"). InnerVolumeSpecName "kube-api-access-d6hbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:37:52 crc kubenswrapper[4998]: I1203 16:37:52.691367 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fb49d67-f5a0-4728-b3e1-dc596955f03a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5fb49d67-f5a0-4728-b3e1-dc596955f03a" (UID: "5fb49d67-f5a0-4728-b3e1-dc596955f03a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:37:52 crc kubenswrapper[4998]: I1203 16:37:52.693414 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fb49d67-f5a0-4728-b3e1-dc596955f03a-inventory" (OuterVolumeSpecName: "inventory") pod "5fb49d67-f5a0-4728-b3e1-dc596955f03a" (UID: "5fb49d67-f5a0-4728-b3e1-dc596955f03a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:37:52 crc kubenswrapper[4998]: I1203 16:37:52.762347 4998 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fb49d67-f5a0-4728-b3e1-dc596955f03a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 16:37:52 crc kubenswrapper[4998]: I1203 16:37:52.762388 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6hbb\" (UniqueName: \"kubernetes.io/projected/5fb49d67-f5a0-4728-b3e1-dc596955f03a-kube-api-access-d6hbb\") on node \"crc\" DevicePath \"\"" Dec 03 16:37:52 crc kubenswrapper[4998]: I1203 16:37:52.762404 4998 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fb49d67-f5a0-4728-b3e1-dc596955f03a-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.124853 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk" event={"ID":"5fb49d67-f5a0-4728-b3e1-dc596955f03a","Type":"ContainerDied","Data":"ef42397c9adb675585fa0c3688e69d1ef06b866ab553220984abadbf6a274fe2"} Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.124898 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef42397c9adb675585fa0c3688e69d1ef06b866ab553220984abadbf6a274fe2" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.124969 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.211217 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd"] Dec 03 16:37:53 crc kubenswrapper[4998]: E1203 16:37:53.212025 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8833f82b-b833-4a08-864f-219787f8f4b8" containerName="registry-server" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.212118 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="8833f82b-b833-4a08-864f-219787f8f4b8" containerName="registry-server" Dec 03 16:37:53 crc kubenswrapper[4998]: E1203 16:37:53.212193 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fb49d67-f5a0-4728-b3e1-dc596955f03a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.212256 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fb49d67-f5a0-4728-b3e1-dc596955f03a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 16:37:53 crc kubenswrapper[4998]: E1203 16:37:53.212339 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8833f82b-b833-4a08-864f-219787f8f4b8" containerName="extract-utilities" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.212407 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="8833f82b-b833-4a08-864f-219787f8f4b8" containerName="extract-utilities" Dec 03 16:37:53 crc kubenswrapper[4998]: E1203 16:37:53.212469 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc03c421-4ea1-46c9-ae67-03af108bf717" containerName="extract-utilities" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.212537 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc03c421-4ea1-46c9-ae67-03af108bf717" containerName="extract-utilities" Dec 03 16:37:53 crc kubenswrapper[4998]: E1203 16:37:53.212629 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8833f82b-b833-4a08-864f-219787f8f4b8" containerName="extract-content" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.212700 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="8833f82b-b833-4a08-864f-219787f8f4b8" containerName="extract-content" Dec 03 16:37:53 crc kubenswrapper[4998]: E1203 16:37:53.212811 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc03c421-4ea1-46c9-ae67-03af108bf717" containerName="registry-server" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.212894 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc03c421-4ea1-46c9-ae67-03af108bf717" containerName="registry-server" Dec 03 16:37:53 crc kubenswrapper[4998]: E1203 16:37:53.212994 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc03c421-4ea1-46c9-ae67-03af108bf717" containerName="extract-content" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.213068 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc03c421-4ea1-46c9-ae67-03af108bf717" containerName="extract-content" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.213441 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc03c421-4ea1-46c9-ae67-03af108bf717" containerName="registry-server" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.213530 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="8833f82b-b833-4a08-864f-219787f8f4b8" containerName="registry-server" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.213616 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fb49d67-f5a0-4728-b3e1-dc596955f03a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.214591 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.219958 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kcj9x" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.220271 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.220532 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.220587 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.223602 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd"] Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.272669 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b407bc5f-1bea-45f6-916a-db593405aab2-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd\" (UID: \"b407bc5f-1bea-45f6-916a-db593405aab2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.272844 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln2rq\" (UniqueName: \"kubernetes.io/projected/b407bc5f-1bea-45f6-916a-db593405aab2-kube-api-access-ln2rq\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd\" (UID: \"b407bc5f-1bea-45f6-916a-db593405aab2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.272897 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b407bc5f-1bea-45f6-916a-db593405aab2-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd\" (UID: \"b407bc5f-1bea-45f6-916a-db593405aab2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.374860 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b407bc5f-1bea-45f6-916a-db593405aab2-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd\" (UID: \"b407bc5f-1bea-45f6-916a-db593405aab2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.375002 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ln2rq\" (UniqueName: \"kubernetes.io/projected/b407bc5f-1bea-45f6-916a-db593405aab2-kube-api-access-ln2rq\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd\" (UID: \"b407bc5f-1bea-45f6-916a-db593405aab2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.375051 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b407bc5f-1bea-45f6-916a-db593405aab2-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd\" (UID: \"b407bc5f-1bea-45f6-916a-db593405aab2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.378736 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b407bc5f-1bea-45f6-916a-db593405aab2-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd\" (UID: \"b407bc5f-1bea-45f6-916a-db593405aab2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.379433 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b407bc5f-1bea-45f6-916a-db593405aab2-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd\" (UID: \"b407bc5f-1bea-45f6-916a-db593405aab2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.391711 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ln2rq\" (UniqueName: \"kubernetes.io/projected/b407bc5f-1bea-45f6-916a-db593405aab2-kube-api-access-ln2rq\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd\" (UID: \"b407bc5f-1bea-45f6-916a-db593405aab2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd" Dec 03 16:37:53 crc kubenswrapper[4998]: I1203 16:37:53.549326 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd" Dec 03 16:37:54 crc kubenswrapper[4998]: I1203 16:37:54.058395 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd"] Dec 03 16:37:54 crc kubenswrapper[4998]: I1203 16:37:54.135788 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd" event={"ID":"b407bc5f-1bea-45f6-916a-db593405aab2","Type":"ContainerStarted","Data":"f612e19a0812e0105cc19e0ebb13a7be29d910b43f90919776aee9507a8ff670"} Dec 03 16:37:55 crc kubenswrapper[4998]: I1203 16:37:55.057129 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-dlfdq"] Dec 03 16:37:55 crc kubenswrapper[4998]: I1203 16:37:55.068475 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-dlfdq"] Dec 03 16:37:55 crc kubenswrapper[4998]: I1203 16:37:55.147449 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd" event={"ID":"b407bc5f-1bea-45f6-916a-db593405aab2","Type":"ContainerStarted","Data":"38783491a2ded6b2dde299ec6104b6763187749120acbe65a8f1fa5b44b63ce2"} Dec 03 16:37:55 crc kubenswrapper[4998]: I1203 16:37:55.167797 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd" podStartSLOduration=1.751150407 podStartE2EDuration="2.167776149s" podCreationTimestamp="2025-12-03 16:37:53 +0000 UTC" firstStartedPulling="2025-12-03 16:37:54.067296617 +0000 UTC m=+2052.678996840" lastFinishedPulling="2025-12-03 16:37:54.483922359 +0000 UTC m=+2053.095622582" observedRunningTime="2025-12-03 16:37:55.162334203 +0000 UTC m=+2053.774034426" watchObservedRunningTime="2025-12-03 16:37:55.167776149 +0000 UTC m=+2053.779476372" Dec 03 16:37:55 crc kubenswrapper[4998]: I1203 16:37:55.691431 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53aaa2e5-6add-4615-9aa1-9af99d56ea04" path="/var/lib/kubelet/pods/53aaa2e5-6add-4615-9aa1-9af99d56ea04/volumes" Dec 03 16:38:00 crc kubenswrapper[4998]: I1203 16:38:00.211914 4998 generic.go:334] "Generic (PLEG): container finished" podID="b407bc5f-1bea-45f6-916a-db593405aab2" containerID="38783491a2ded6b2dde299ec6104b6763187749120acbe65a8f1fa5b44b63ce2" exitCode=0 Dec 03 16:38:00 crc kubenswrapper[4998]: I1203 16:38:00.212000 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd" event={"ID":"b407bc5f-1bea-45f6-916a-db593405aab2","Type":"ContainerDied","Data":"38783491a2ded6b2dde299ec6104b6763187749120acbe65a8f1fa5b44b63ce2"} Dec 03 16:38:00 crc kubenswrapper[4998]: I1203 16:38:00.286977 4998 scope.go:117] "RemoveContainer" containerID="ae7a2f32fd866e6275ac3717266de1857c58450ea3f2547815e726e96f085b84" Dec 03 16:38:00 crc kubenswrapper[4998]: I1203 16:38:00.337153 4998 scope.go:117] "RemoveContainer" containerID="d7a9fa9786f0425c5472b4b8e59b0cb7bf46e79182787c3b5551ee7e4116bc06" Dec 03 16:38:00 crc kubenswrapper[4998]: I1203 16:38:00.375833 4998 scope.go:117] "RemoveContainer" containerID="e9f429e5e1f1bd89d2bb9cca65fdc0eebb511a8d993f432a17e22b665be35a4c" Dec 03 16:38:01 crc kubenswrapper[4998]: I1203 16:38:01.648152 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd" Dec 03 16:38:01 crc kubenswrapper[4998]: I1203 16:38:01.762408 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b407bc5f-1bea-45f6-916a-db593405aab2-ssh-key\") pod \"b407bc5f-1bea-45f6-916a-db593405aab2\" (UID: \"b407bc5f-1bea-45f6-916a-db593405aab2\") " Dec 03 16:38:01 crc kubenswrapper[4998]: I1203 16:38:01.762547 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b407bc5f-1bea-45f6-916a-db593405aab2-inventory\") pod \"b407bc5f-1bea-45f6-916a-db593405aab2\" (UID: \"b407bc5f-1bea-45f6-916a-db593405aab2\") " Dec 03 16:38:01 crc kubenswrapper[4998]: I1203 16:38:01.762725 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ln2rq\" (UniqueName: \"kubernetes.io/projected/b407bc5f-1bea-45f6-916a-db593405aab2-kube-api-access-ln2rq\") pod \"b407bc5f-1bea-45f6-916a-db593405aab2\" (UID: \"b407bc5f-1bea-45f6-916a-db593405aab2\") " Dec 03 16:38:01 crc kubenswrapper[4998]: I1203 16:38:01.769380 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b407bc5f-1bea-45f6-916a-db593405aab2-kube-api-access-ln2rq" (OuterVolumeSpecName: "kube-api-access-ln2rq") pod "b407bc5f-1bea-45f6-916a-db593405aab2" (UID: "b407bc5f-1bea-45f6-916a-db593405aab2"). InnerVolumeSpecName "kube-api-access-ln2rq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:38:01 crc kubenswrapper[4998]: I1203 16:38:01.791958 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b407bc5f-1bea-45f6-916a-db593405aab2-inventory" (OuterVolumeSpecName: "inventory") pod "b407bc5f-1bea-45f6-916a-db593405aab2" (UID: "b407bc5f-1bea-45f6-916a-db593405aab2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:38:01 crc kubenswrapper[4998]: I1203 16:38:01.806176 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b407bc5f-1bea-45f6-916a-db593405aab2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b407bc5f-1bea-45f6-916a-db593405aab2" (UID: "b407bc5f-1bea-45f6-916a-db593405aab2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:38:01 crc kubenswrapper[4998]: I1203 16:38:01.865811 4998 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b407bc5f-1bea-45f6-916a-db593405aab2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 16:38:01 crc kubenswrapper[4998]: I1203 16:38:01.865885 4998 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b407bc5f-1bea-45f6-916a-db593405aab2-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 16:38:01 crc kubenswrapper[4998]: I1203 16:38:01.865901 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ln2rq\" (UniqueName: \"kubernetes.io/projected/b407bc5f-1bea-45f6-916a-db593405aab2-kube-api-access-ln2rq\") on node \"crc\" DevicePath \"\"" Dec 03 16:38:02 crc kubenswrapper[4998]: I1203 16:38:02.236452 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd" event={"ID":"b407bc5f-1bea-45f6-916a-db593405aab2","Type":"ContainerDied","Data":"f612e19a0812e0105cc19e0ebb13a7be29d910b43f90919776aee9507a8ff670"} Dec 03 16:38:02 crc kubenswrapper[4998]: I1203 16:38:02.236503 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f612e19a0812e0105cc19e0ebb13a7be29d910b43f90919776aee9507a8ff670" Dec 03 16:38:02 crc kubenswrapper[4998]: I1203 16:38:02.236583 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd" Dec 03 16:38:02 crc kubenswrapper[4998]: I1203 16:38:02.332214 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-g8wgf"] Dec 03 16:38:02 crc kubenswrapper[4998]: E1203 16:38:02.332701 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b407bc5f-1bea-45f6-916a-db593405aab2" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 16:38:02 crc kubenswrapper[4998]: I1203 16:38:02.332727 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="b407bc5f-1bea-45f6-916a-db593405aab2" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 16:38:02 crc kubenswrapper[4998]: I1203 16:38:02.333009 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="b407bc5f-1bea-45f6-916a-db593405aab2" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 16:38:02 crc kubenswrapper[4998]: I1203 16:38:02.333842 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g8wgf" Dec 03 16:38:02 crc kubenswrapper[4998]: I1203 16:38:02.338307 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 16:38:02 crc kubenswrapper[4998]: I1203 16:38:02.338472 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kcj9x" Dec 03 16:38:02 crc kubenswrapper[4998]: I1203 16:38:02.338587 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 16:38:02 crc kubenswrapper[4998]: I1203 16:38:02.338805 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 16:38:02 crc kubenswrapper[4998]: I1203 16:38:02.346949 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-g8wgf"] Dec 03 16:38:02 crc kubenswrapper[4998]: I1203 16:38:02.378844 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f58e5a8-bad5-4d0e-ae33-8e666718bf0a-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g8wgf\" (UID: \"6f58e5a8-bad5-4d0e-ae33-8e666718bf0a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g8wgf" Dec 03 16:38:02 crc kubenswrapper[4998]: I1203 16:38:02.378991 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lgjk\" (UniqueName: \"kubernetes.io/projected/6f58e5a8-bad5-4d0e-ae33-8e666718bf0a-kube-api-access-5lgjk\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g8wgf\" (UID: \"6f58e5a8-bad5-4d0e-ae33-8e666718bf0a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g8wgf" Dec 03 16:38:02 crc kubenswrapper[4998]: I1203 16:38:02.379315 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f58e5a8-bad5-4d0e-ae33-8e666718bf0a-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g8wgf\" (UID: \"6f58e5a8-bad5-4d0e-ae33-8e666718bf0a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g8wgf" Dec 03 16:38:02 crc kubenswrapper[4998]: I1203 16:38:02.481287 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lgjk\" (UniqueName: \"kubernetes.io/projected/6f58e5a8-bad5-4d0e-ae33-8e666718bf0a-kube-api-access-5lgjk\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g8wgf\" (UID: \"6f58e5a8-bad5-4d0e-ae33-8e666718bf0a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g8wgf" Dec 03 16:38:02 crc kubenswrapper[4998]: I1203 16:38:02.481393 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f58e5a8-bad5-4d0e-ae33-8e666718bf0a-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g8wgf\" (UID: \"6f58e5a8-bad5-4d0e-ae33-8e666718bf0a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g8wgf" Dec 03 16:38:02 crc kubenswrapper[4998]: I1203 16:38:02.481472 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f58e5a8-bad5-4d0e-ae33-8e666718bf0a-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g8wgf\" (UID: \"6f58e5a8-bad5-4d0e-ae33-8e666718bf0a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g8wgf" Dec 03 16:38:02 crc kubenswrapper[4998]: I1203 16:38:02.486965 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f58e5a8-bad5-4d0e-ae33-8e666718bf0a-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g8wgf\" (UID: \"6f58e5a8-bad5-4d0e-ae33-8e666718bf0a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g8wgf" Dec 03 16:38:02 crc kubenswrapper[4998]: I1203 16:38:02.488089 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f58e5a8-bad5-4d0e-ae33-8e666718bf0a-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g8wgf\" (UID: \"6f58e5a8-bad5-4d0e-ae33-8e666718bf0a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g8wgf" Dec 03 16:38:02 crc kubenswrapper[4998]: I1203 16:38:02.499999 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lgjk\" (UniqueName: \"kubernetes.io/projected/6f58e5a8-bad5-4d0e-ae33-8e666718bf0a-kube-api-access-5lgjk\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g8wgf\" (UID: \"6f58e5a8-bad5-4d0e-ae33-8e666718bf0a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g8wgf" Dec 03 16:38:02 crc kubenswrapper[4998]: I1203 16:38:02.661793 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g8wgf" Dec 03 16:38:03 crc kubenswrapper[4998]: I1203 16:38:03.256096 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-g8wgf"] Dec 03 16:38:04 crc kubenswrapper[4998]: I1203 16:38:04.255827 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g8wgf" event={"ID":"6f58e5a8-bad5-4d0e-ae33-8e666718bf0a","Type":"ContainerStarted","Data":"f30867bfc715fbb0c5823af2a5db795f677d7ecbb75a1f53b14f4773ab10137d"} Dec 03 16:38:04 crc kubenswrapper[4998]: I1203 16:38:04.256251 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g8wgf" event={"ID":"6f58e5a8-bad5-4d0e-ae33-8e666718bf0a","Type":"ContainerStarted","Data":"527296340cb2063e98ba88fc6259ce437b592f81c92a63a20c6d7ac461adfbf6"} Dec 03 16:38:04 crc kubenswrapper[4998]: I1203 16:38:04.280901 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g8wgf" podStartSLOduration=1.801459079 podStartE2EDuration="2.280880415s" podCreationTimestamp="2025-12-03 16:38:02 +0000 UTC" firstStartedPulling="2025-12-03 16:38:03.269660798 +0000 UTC m=+2061.881361011" lastFinishedPulling="2025-12-03 16:38:03.749082124 +0000 UTC m=+2062.360782347" observedRunningTime="2025-12-03 16:38:04.270100197 +0000 UTC m=+2062.881800440" watchObservedRunningTime="2025-12-03 16:38:04.280880415 +0000 UTC m=+2062.892580638" Dec 03 16:38:49 crc kubenswrapper[4998]: I1203 16:38:49.741669 4998 generic.go:334] "Generic (PLEG): container finished" podID="6f58e5a8-bad5-4d0e-ae33-8e666718bf0a" containerID="f30867bfc715fbb0c5823af2a5db795f677d7ecbb75a1f53b14f4773ab10137d" exitCode=0 Dec 03 16:38:49 crc kubenswrapper[4998]: I1203 16:38:49.741807 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g8wgf" event={"ID":"6f58e5a8-bad5-4d0e-ae33-8e666718bf0a","Type":"ContainerDied","Data":"f30867bfc715fbb0c5823af2a5db795f677d7ecbb75a1f53b14f4773ab10137d"} Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.276563 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g8wgf" Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.375010 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lgjk\" (UniqueName: \"kubernetes.io/projected/6f58e5a8-bad5-4d0e-ae33-8e666718bf0a-kube-api-access-5lgjk\") pod \"6f58e5a8-bad5-4d0e-ae33-8e666718bf0a\" (UID: \"6f58e5a8-bad5-4d0e-ae33-8e666718bf0a\") " Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.375074 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f58e5a8-bad5-4d0e-ae33-8e666718bf0a-inventory\") pod \"6f58e5a8-bad5-4d0e-ae33-8e666718bf0a\" (UID: \"6f58e5a8-bad5-4d0e-ae33-8e666718bf0a\") " Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.375120 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f58e5a8-bad5-4d0e-ae33-8e666718bf0a-ssh-key\") pod \"6f58e5a8-bad5-4d0e-ae33-8e666718bf0a\" (UID: \"6f58e5a8-bad5-4d0e-ae33-8e666718bf0a\") " Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.379905 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f58e5a8-bad5-4d0e-ae33-8e666718bf0a-kube-api-access-5lgjk" (OuterVolumeSpecName: "kube-api-access-5lgjk") pod "6f58e5a8-bad5-4d0e-ae33-8e666718bf0a" (UID: "6f58e5a8-bad5-4d0e-ae33-8e666718bf0a"). InnerVolumeSpecName "kube-api-access-5lgjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.408366 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f58e5a8-bad5-4d0e-ae33-8e666718bf0a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6f58e5a8-bad5-4d0e-ae33-8e666718bf0a" (UID: "6f58e5a8-bad5-4d0e-ae33-8e666718bf0a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.414424 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f58e5a8-bad5-4d0e-ae33-8e666718bf0a-inventory" (OuterVolumeSpecName: "inventory") pod "6f58e5a8-bad5-4d0e-ae33-8e666718bf0a" (UID: "6f58e5a8-bad5-4d0e-ae33-8e666718bf0a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.477363 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lgjk\" (UniqueName: \"kubernetes.io/projected/6f58e5a8-bad5-4d0e-ae33-8e666718bf0a-kube-api-access-5lgjk\") on node \"crc\" DevicePath \"\"" Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.477401 4998 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f58e5a8-bad5-4d0e-ae33-8e666718bf0a-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.477414 4998 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f58e5a8-bad5-4d0e-ae33-8e666718bf0a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.766344 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g8wgf" event={"ID":"6f58e5a8-bad5-4d0e-ae33-8e666718bf0a","Type":"ContainerDied","Data":"527296340cb2063e98ba88fc6259ce437b592f81c92a63a20c6d7ac461adfbf6"} Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.766390 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="527296340cb2063e98ba88fc6259ce437b592f81c92a63a20c6d7ac461adfbf6" Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.766673 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g8wgf" Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.865342 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v57vv"] Dec 03 16:38:51 crc kubenswrapper[4998]: E1203 16:38:51.866118 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f58e5a8-bad5-4d0e-ae33-8e666718bf0a" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.866137 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f58e5a8-bad5-4d0e-ae33-8e666718bf0a" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.866351 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f58e5a8-bad5-4d0e-ae33-8e666718bf0a" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.867116 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v57vv" Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.869881 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.869927 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.871514 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.871788 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kcj9x" Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.878863 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v57vv"] Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.884891 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bd61b3c-44d3-46c3-9e4e-094937642ef8-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v57vv\" (UID: \"4bd61b3c-44d3-46c3-9e4e-094937642ef8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v57vv" Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.884935 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bd61b3c-44d3-46c3-9e4e-094937642ef8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v57vv\" (UID: \"4bd61b3c-44d3-46c3-9e4e-094937642ef8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v57vv" Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.885099 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vkps\" (UniqueName: \"kubernetes.io/projected/4bd61b3c-44d3-46c3-9e4e-094937642ef8-kube-api-access-7vkps\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v57vv\" (UID: \"4bd61b3c-44d3-46c3-9e4e-094937642ef8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v57vv" Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.987599 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vkps\" (UniqueName: \"kubernetes.io/projected/4bd61b3c-44d3-46c3-9e4e-094937642ef8-kube-api-access-7vkps\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v57vv\" (UID: \"4bd61b3c-44d3-46c3-9e4e-094937642ef8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v57vv" Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.987715 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bd61b3c-44d3-46c3-9e4e-094937642ef8-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v57vv\" (UID: \"4bd61b3c-44d3-46c3-9e4e-094937642ef8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v57vv" Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.987781 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bd61b3c-44d3-46c3-9e4e-094937642ef8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v57vv\" (UID: \"4bd61b3c-44d3-46c3-9e4e-094937642ef8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v57vv" Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.992388 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bd61b3c-44d3-46c3-9e4e-094937642ef8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v57vv\" (UID: \"4bd61b3c-44d3-46c3-9e4e-094937642ef8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v57vv" Dec 03 16:38:51 crc kubenswrapper[4998]: I1203 16:38:51.997836 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bd61b3c-44d3-46c3-9e4e-094937642ef8-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v57vv\" (UID: \"4bd61b3c-44d3-46c3-9e4e-094937642ef8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v57vv" Dec 03 16:38:52 crc kubenswrapper[4998]: I1203 16:38:52.011732 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vkps\" (UniqueName: \"kubernetes.io/projected/4bd61b3c-44d3-46c3-9e4e-094937642ef8-kube-api-access-7vkps\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-v57vv\" (UID: \"4bd61b3c-44d3-46c3-9e4e-094937642ef8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v57vv" Dec 03 16:38:52 crc kubenswrapper[4998]: I1203 16:38:52.184051 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v57vv" Dec 03 16:38:52 crc kubenswrapper[4998]: I1203 16:38:52.737953 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v57vv"] Dec 03 16:38:52 crc kubenswrapper[4998]: I1203 16:38:52.774000 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v57vv" event={"ID":"4bd61b3c-44d3-46c3-9e4e-094937642ef8","Type":"ContainerStarted","Data":"178b639a9aa4f14d1ae078dbc71029a0c503e5a977ab79ae9983eefb0dcaccc2"} Dec 03 16:38:53 crc kubenswrapper[4998]: I1203 16:38:53.064887 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p5mxv"] Dec 03 16:38:53 crc kubenswrapper[4998]: I1203 16:38:53.067984 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p5mxv" Dec 03 16:38:53 crc kubenswrapper[4998]: I1203 16:38:53.079795 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p5mxv"] Dec 03 16:38:53 crc kubenswrapper[4998]: I1203 16:38:53.236028 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/145328e4-60b9-49c0-8e9a-36007a5dc2c3-catalog-content\") pod \"certified-operators-p5mxv\" (UID: \"145328e4-60b9-49c0-8e9a-36007a5dc2c3\") " pod="openshift-marketplace/certified-operators-p5mxv" Dec 03 16:38:53 crc kubenswrapper[4998]: I1203 16:38:53.237218 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/145328e4-60b9-49c0-8e9a-36007a5dc2c3-utilities\") pod \"certified-operators-p5mxv\" (UID: \"145328e4-60b9-49c0-8e9a-36007a5dc2c3\") " pod="openshift-marketplace/certified-operators-p5mxv" Dec 03 16:38:53 crc kubenswrapper[4998]: I1203 16:38:53.237402 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvdkq\" (UniqueName: \"kubernetes.io/projected/145328e4-60b9-49c0-8e9a-36007a5dc2c3-kube-api-access-zvdkq\") pod \"certified-operators-p5mxv\" (UID: \"145328e4-60b9-49c0-8e9a-36007a5dc2c3\") " pod="openshift-marketplace/certified-operators-p5mxv" Dec 03 16:38:53 crc kubenswrapper[4998]: I1203 16:38:53.340034 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvdkq\" (UniqueName: \"kubernetes.io/projected/145328e4-60b9-49c0-8e9a-36007a5dc2c3-kube-api-access-zvdkq\") pod \"certified-operators-p5mxv\" (UID: \"145328e4-60b9-49c0-8e9a-36007a5dc2c3\") " pod="openshift-marketplace/certified-operators-p5mxv" Dec 03 16:38:53 crc kubenswrapper[4998]: I1203 16:38:53.340436 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/145328e4-60b9-49c0-8e9a-36007a5dc2c3-catalog-content\") pod \"certified-operators-p5mxv\" (UID: \"145328e4-60b9-49c0-8e9a-36007a5dc2c3\") " pod="openshift-marketplace/certified-operators-p5mxv" Dec 03 16:38:53 crc kubenswrapper[4998]: I1203 16:38:53.340521 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/145328e4-60b9-49c0-8e9a-36007a5dc2c3-utilities\") pod \"certified-operators-p5mxv\" (UID: \"145328e4-60b9-49c0-8e9a-36007a5dc2c3\") " pod="openshift-marketplace/certified-operators-p5mxv" Dec 03 16:38:53 crc kubenswrapper[4998]: I1203 16:38:53.340971 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/145328e4-60b9-49c0-8e9a-36007a5dc2c3-utilities\") pod \"certified-operators-p5mxv\" (UID: \"145328e4-60b9-49c0-8e9a-36007a5dc2c3\") " pod="openshift-marketplace/certified-operators-p5mxv" Dec 03 16:38:53 crc kubenswrapper[4998]: I1203 16:38:53.340971 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/145328e4-60b9-49c0-8e9a-36007a5dc2c3-catalog-content\") pod \"certified-operators-p5mxv\" (UID: \"145328e4-60b9-49c0-8e9a-36007a5dc2c3\") " pod="openshift-marketplace/certified-operators-p5mxv" Dec 03 16:38:53 crc kubenswrapper[4998]: I1203 16:38:53.360271 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvdkq\" (UniqueName: \"kubernetes.io/projected/145328e4-60b9-49c0-8e9a-36007a5dc2c3-kube-api-access-zvdkq\") pod \"certified-operators-p5mxv\" (UID: \"145328e4-60b9-49c0-8e9a-36007a5dc2c3\") " pod="openshift-marketplace/certified-operators-p5mxv" Dec 03 16:38:53 crc kubenswrapper[4998]: I1203 16:38:53.401467 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p5mxv" Dec 03 16:38:53 crc kubenswrapper[4998]: I1203 16:38:53.790046 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v57vv" event={"ID":"4bd61b3c-44d3-46c3-9e4e-094937642ef8","Type":"ContainerStarted","Data":"e22c1ea258c62b67dc4b999e00860f7d7d0c3cdea8a73d94a4a2b991108b5e87"} Dec 03 16:38:53 crc kubenswrapper[4998]: I1203 16:38:53.817815 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v57vv" podStartSLOduration=2.306974821 podStartE2EDuration="2.817790649s" podCreationTimestamp="2025-12-03 16:38:51 +0000 UTC" firstStartedPulling="2025-12-03 16:38:52.741838989 +0000 UTC m=+2111.353539212" lastFinishedPulling="2025-12-03 16:38:53.252654817 +0000 UTC m=+2111.864355040" observedRunningTime="2025-12-03 16:38:53.809582534 +0000 UTC m=+2112.421282767" watchObservedRunningTime="2025-12-03 16:38:53.817790649 +0000 UTC m=+2112.429490892" Dec 03 16:38:54 crc kubenswrapper[4998]: W1203 16:38:54.145035 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod145328e4_60b9_49c0_8e9a_36007a5dc2c3.slice/crio-93fe1d3f985e2ce46dd0c5b07bf2facae678d16087f3b575c59d672d7332c3b8 WatchSource:0}: Error finding container 93fe1d3f985e2ce46dd0c5b07bf2facae678d16087f3b575c59d672d7332c3b8: Status 404 returned error can't find the container with id 93fe1d3f985e2ce46dd0c5b07bf2facae678d16087f3b575c59d672d7332c3b8 Dec 03 16:38:54 crc kubenswrapper[4998]: I1203 16:38:54.147299 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p5mxv"] Dec 03 16:38:54 crc kubenswrapper[4998]: I1203 16:38:54.805363 4998 generic.go:334] "Generic (PLEG): container finished" podID="145328e4-60b9-49c0-8e9a-36007a5dc2c3" containerID="cd31398601291c738a942c4de8e2d58c03cf01fb165f8d9c0580dcf65de33a50" exitCode=0 Dec 03 16:38:54 crc kubenswrapper[4998]: I1203 16:38:54.805470 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5mxv" event={"ID":"145328e4-60b9-49c0-8e9a-36007a5dc2c3","Type":"ContainerDied","Data":"cd31398601291c738a942c4de8e2d58c03cf01fb165f8d9c0580dcf65de33a50"} Dec 03 16:38:54 crc kubenswrapper[4998]: I1203 16:38:54.805531 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5mxv" event={"ID":"145328e4-60b9-49c0-8e9a-36007a5dc2c3","Type":"ContainerStarted","Data":"93fe1d3f985e2ce46dd0c5b07bf2facae678d16087f3b575c59d672d7332c3b8"} Dec 03 16:38:55 crc kubenswrapper[4998]: I1203 16:38:55.818934 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5mxv" event={"ID":"145328e4-60b9-49c0-8e9a-36007a5dc2c3","Type":"ContainerStarted","Data":"57b28587bea82ed8f5d1a47bc1c18b297a13089cba4974b80402e9dbef97101c"} Dec 03 16:38:56 crc kubenswrapper[4998]: I1203 16:38:56.831046 4998 generic.go:334] "Generic (PLEG): container finished" podID="145328e4-60b9-49c0-8e9a-36007a5dc2c3" containerID="57b28587bea82ed8f5d1a47bc1c18b297a13089cba4974b80402e9dbef97101c" exitCode=0 Dec 03 16:38:56 crc kubenswrapper[4998]: I1203 16:38:56.831147 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5mxv" event={"ID":"145328e4-60b9-49c0-8e9a-36007a5dc2c3","Type":"ContainerDied","Data":"57b28587bea82ed8f5d1a47bc1c18b297a13089cba4974b80402e9dbef97101c"} Dec 03 16:38:57 crc kubenswrapper[4998]: I1203 16:38:57.111626 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:38:57 crc kubenswrapper[4998]: I1203 16:38:57.111682 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:38:58 crc kubenswrapper[4998]: I1203 16:38:58.856974 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5mxv" event={"ID":"145328e4-60b9-49c0-8e9a-36007a5dc2c3","Type":"ContainerStarted","Data":"52d0a974098ce0311502303359bf9204dc7b219a35e4c7275fdec0c465da7d4f"} Dec 03 16:38:58 crc kubenswrapper[4998]: I1203 16:38:58.892201 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p5mxv" podStartSLOduration=2.682057423 podStartE2EDuration="5.892160441s" podCreationTimestamp="2025-12-03 16:38:53 +0000 UTC" firstStartedPulling="2025-12-03 16:38:54.817210282 +0000 UTC m=+2113.428910495" lastFinishedPulling="2025-12-03 16:38:58.02731329 +0000 UTC m=+2116.639013513" observedRunningTime="2025-12-03 16:38:58.882250024 +0000 UTC m=+2117.493950287" watchObservedRunningTime="2025-12-03 16:38:58.892160441 +0000 UTC m=+2117.503860654" Dec 03 16:39:03 crc kubenswrapper[4998]: I1203 16:39:03.401822 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p5mxv" Dec 03 16:39:03 crc kubenswrapper[4998]: I1203 16:39:03.403138 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p5mxv" Dec 03 16:39:03 crc kubenswrapper[4998]: I1203 16:39:03.459430 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p5mxv" Dec 03 16:39:03 crc kubenswrapper[4998]: I1203 16:39:03.979497 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p5mxv" Dec 03 16:39:04 crc kubenswrapper[4998]: I1203 16:39:04.042701 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p5mxv"] Dec 03 16:39:05 crc kubenswrapper[4998]: I1203 16:39:05.926634 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p5mxv" podUID="145328e4-60b9-49c0-8e9a-36007a5dc2c3" containerName="registry-server" containerID="cri-o://52d0a974098ce0311502303359bf9204dc7b219a35e4c7275fdec0c465da7d4f" gracePeriod=2 Dec 03 16:39:06 crc kubenswrapper[4998]: E1203 16:39:06.104786 4998 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod145328e4_60b9_49c0_8e9a_36007a5dc2c3.slice/crio-52d0a974098ce0311502303359bf9204dc7b219a35e4c7275fdec0c465da7d4f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod145328e4_60b9_49c0_8e9a_36007a5dc2c3.slice/crio-conmon-52d0a974098ce0311502303359bf9204dc7b219a35e4c7275fdec0c465da7d4f.scope\": RecentStats: unable to find data in memory cache]" Dec 03 16:39:06 crc kubenswrapper[4998]: I1203 16:39:06.475505 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p5mxv" Dec 03 16:39:06 crc kubenswrapper[4998]: I1203 16:39:06.630939 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/145328e4-60b9-49c0-8e9a-36007a5dc2c3-catalog-content\") pod \"145328e4-60b9-49c0-8e9a-36007a5dc2c3\" (UID: \"145328e4-60b9-49c0-8e9a-36007a5dc2c3\") " Dec 03 16:39:06 crc kubenswrapper[4998]: I1203 16:39:06.630983 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvdkq\" (UniqueName: \"kubernetes.io/projected/145328e4-60b9-49c0-8e9a-36007a5dc2c3-kube-api-access-zvdkq\") pod \"145328e4-60b9-49c0-8e9a-36007a5dc2c3\" (UID: \"145328e4-60b9-49c0-8e9a-36007a5dc2c3\") " Dec 03 16:39:06 crc kubenswrapper[4998]: I1203 16:39:06.631139 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/145328e4-60b9-49c0-8e9a-36007a5dc2c3-utilities\") pod \"145328e4-60b9-49c0-8e9a-36007a5dc2c3\" (UID: \"145328e4-60b9-49c0-8e9a-36007a5dc2c3\") " Dec 03 16:39:06 crc kubenswrapper[4998]: I1203 16:39:06.632049 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/145328e4-60b9-49c0-8e9a-36007a5dc2c3-utilities" (OuterVolumeSpecName: "utilities") pod "145328e4-60b9-49c0-8e9a-36007a5dc2c3" (UID: "145328e4-60b9-49c0-8e9a-36007a5dc2c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:39:06 crc kubenswrapper[4998]: I1203 16:39:06.641111 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/145328e4-60b9-49c0-8e9a-36007a5dc2c3-kube-api-access-zvdkq" (OuterVolumeSpecName: "kube-api-access-zvdkq") pod "145328e4-60b9-49c0-8e9a-36007a5dc2c3" (UID: "145328e4-60b9-49c0-8e9a-36007a5dc2c3"). InnerVolumeSpecName "kube-api-access-zvdkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:39:06 crc kubenswrapper[4998]: I1203 16:39:06.715630 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/145328e4-60b9-49c0-8e9a-36007a5dc2c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "145328e4-60b9-49c0-8e9a-36007a5dc2c3" (UID: "145328e4-60b9-49c0-8e9a-36007a5dc2c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:39:06 crc kubenswrapper[4998]: I1203 16:39:06.733798 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvdkq\" (UniqueName: \"kubernetes.io/projected/145328e4-60b9-49c0-8e9a-36007a5dc2c3-kube-api-access-zvdkq\") on node \"crc\" DevicePath \"\"" Dec 03 16:39:06 crc kubenswrapper[4998]: I1203 16:39:06.733836 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/145328e4-60b9-49c0-8e9a-36007a5dc2c3-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:39:06 crc kubenswrapper[4998]: I1203 16:39:06.733849 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/145328e4-60b9-49c0-8e9a-36007a5dc2c3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:39:06 crc kubenswrapper[4998]: I1203 16:39:06.937478 4998 generic.go:334] "Generic (PLEG): container finished" podID="145328e4-60b9-49c0-8e9a-36007a5dc2c3" containerID="52d0a974098ce0311502303359bf9204dc7b219a35e4c7275fdec0c465da7d4f" exitCode=0 Dec 03 16:39:06 crc kubenswrapper[4998]: I1203 16:39:06.937584 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5mxv" event={"ID":"145328e4-60b9-49c0-8e9a-36007a5dc2c3","Type":"ContainerDied","Data":"52d0a974098ce0311502303359bf9204dc7b219a35e4c7275fdec0c465da7d4f"} Dec 03 16:39:06 crc kubenswrapper[4998]: I1203 16:39:06.937853 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5mxv" event={"ID":"145328e4-60b9-49c0-8e9a-36007a5dc2c3","Type":"ContainerDied","Data":"93fe1d3f985e2ce46dd0c5b07bf2facae678d16087f3b575c59d672d7332c3b8"} Dec 03 16:39:06 crc kubenswrapper[4998]: I1203 16:39:06.937883 4998 scope.go:117] "RemoveContainer" containerID="52d0a974098ce0311502303359bf9204dc7b219a35e4c7275fdec0c465da7d4f" Dec 03 16:39:06 crc kubenswrapper[4998]: I1203 16:39:06.937595 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p5mxv" Dec 03 16:39:06 crc kubenswrapper[4998]: I1203 16:39:06.959270 4998 scope.go:117] "RemoveContainer" containerID="57b28587bea82ed8f5d1a47bc1c18b297a13089cba4974b80402e9dbef97101c" Dec 03 16:39:06 crc kubenswrapper[4998]: I1203 16:39:06.983343 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p5mxv"] Dec 03 16:39:06 crc kubenswrapper[4998]: I1203 16:39:06.992386 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p5mxv"] Dec 03 16:39:06 crc kubenswrapper[4998]: I1203 16:39:06.994340 4998 scope.go:117] "RemoveContainer" containerID="cd31398601291c738a942c4de8e2d58c03cf01fb165f8d9c0580dcf65de33a50" Dec 03 16:39:07 crc kubenswrapper[4998]: I1203 16:39:07.028831 4998 scope.go:117] "RemoveContainer" containerID="52d0a974098ce0311502303359bf9204dc7b219a35e4c7275fdec0c465da7d4f" Dec 03 16:39:07 crc kubenswrapper[4998]: E1203 16:39:07.029271 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52d0a974098ce0311502303359bf9204dc7b219a35e4c7275fdec0c465da7d4f\": container with ID starting with 52d0a974098ce0311502303359bf9204dc7b219a35e4c7275fdec0c465da7d4f not found: ID does not exist" containerID="52d0a974098ce0311502303359bf9204dc7b219a35e4c7275fdec0c465da7d4f" Dec 03 16:39:07 crc kubenswrapper[4998]: I1203 16:39:07.029303 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52d0a974098ce0311502303359bf9204dc7b219a35e4c7275fdec0c465da7d4f"} err="failed to get container status \"52d0a974098ce0311502303359bf9204dc7b219a35e4c7275fdec0c465da7d4f\": rpc error: code = NotFound desc = could not find container \"52d0a974098ce0311502303359bf9204dc7b219a35e4c7275fdec0c465da7d4f\": container with ID starting with 52d0a974098ce0311502303359bf9204dc7b219a35e4c7275fdec0c465da7d4f not found: ID does not exist" Dec 03 16:39:07 crc kubenswrapper[4998]: I1203 16:39:07.029325 4998 scope.go:117] "RemoveContainer" containerID="57b28587bea82ed8f5d1a47bc1c18b297a13089cba4974b80402e9dbef97101c" Dec 03 16:39:07 crc kubenswrapper[4998]: E1203 16:39:07.029802 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57b28587bea82ed8f5d1a47bc1c18b297a13089cba4974b80402e9dbef97101c\": container with ID starting with 57b28587bea82ed8f5d1a47bc1c18b297a13089cba4974b80402e9dbef97101c not found: ID does not exist" containerID="57b28587bea82ed8f5d1a47bc1c18b297a13089cba4974b80402e9dbef97101c" Dec 03 16:39:07 crc kubenswrapper[4998]: I1203 16:39:07.029828 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57b28587bea82ed8f5d1a47bc1c18b297a13089cba4974b80402e9dbef97101c"} err="failed to get container status \"57b28587bea82ed8f5d1a47bc1c18b297a13089cba4974b80402e9dbef97101c\": rpc error: code = NotFound desc = could not find container \"57b28587bea82ed8f5d1a47bc1c18b297a13089cba4974b80402e9dbef97101c\": container with ID starting with 57b28587bea82ed8f5d1a47bc1c18b297a13089cba4974b80402e9dbef97101c not found: ID does not exist" Dec 03 16:39:07 crc kubenswrapper[4998]: I1203 16:39:07.029848 4998 scope.go:117] "RemoveContainer" containerID="cd31398601291c738a942c4de8e2d58c03cf01fb165f8d9c0580dcf65de33a50" Dec 03 16:39:07 crc kubenswrapper[4998]: E1203 16:39:07.030142 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd31398601291c738a942c4de8e2d58c03cf01fb165f8d9c0580dcf65de33a50\": container with ID starting with cd31398601291c738a942c4de8e2d58c03cf01fb165f8d9c0580dcf65de33a50 not found: ID does not exist" containerID="cd31398601291c738a942c4de8e2d58c03cf01fb165f8d9c0580dcf65de33a50" Dec 03 16:39:07 crc kubenswrapper[4998]: I1203 16:39:07.030159 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd31398601291c738a942c4de8e2d58c03cf01fb165f8d9c0580dcf65de33a50"} err="failed to get container status \"cd31398601291c738a942c4de8e2d58c03cf01fb165f8d9c0580dcf65de33a50\": rpc error: code = NotFound desc = could not find container \"cd31398601291c738a942c4de8e2d58c03cf01fb165f8d9c0580dcf65de33a50\": container with ID starting with cd31398601291c738a942c4de8e2d58c03cf01fb165f8d9c0580dcf65de33a50 not found: ID does not exist" Dec 03 16:39:07 crc kubenswrapper[4998]: I1203 16:39:07.690130 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="145328e4-60b9-49c0-8e9a-36007a5dc2c3" path="/var/lib/kubelet/pods/145328e4-60b9-49c0-8e9a-36007a5dc2c3/volumes" Dec 03 16:39:27 crc kubenswrapper[4998]: I1203 16:39:27.111194 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:39:27 crc kubenswrapper[4998]: I1203 16:39:27.111717 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:39:49 crc kubenswrapper[4998]: I1203 16:39:49.375617 4998 generic.go:334] "Generic (PLEG): container finished" podID="4bd61b3c-44d3-46c3-9e4e-094937642ef8" containerID="e22c1ea258c62b67dc4b999e00860f7d7d0c3cdea8a73d94a4a2b991108b5e87" exitCode=0 Dec 03 16:39:49 crc kubenswrapper[4998]: I1203 16:39:49.375722 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v57vv" event={"ID":"4bd61b3c-44d3-46c3-9e4e-094937642ef8","Type":"ContainerDied","Data":"e22c1ea258c62b67dc4b999e00860f7d7d0c3cdea8a73d94a4a2b991108b5e87"} Dec 03 16:39:50 crc kubenswrapper[4998]: I1203 16:39:50.796253 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v57vv" Dec 03 16:39:50 crc kubenswrapper[4998]: I1203 16:39:50.871217 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bd61b3c-44d3-46c3-9e4e-094937642ef8-ssh-key\") pod \"4bd61b3c-44d3-46c3-9e4e-094937642ef8\" (UID: \"4bd61b3c-44d3-46c3-9e4e-094937642ef8\") " Dec 03 16:39:50 crc kubenswrapper[4998]: I1203 16:39:50.871399 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vkps\" (UniqueName: \"kubernetes.io/projected/4bd61b3c-44d3-46c3-9e4e-094937642ef8-kube-api-access-7vkps\") pod \"4bd61b3c-44d3-46c3-9e4e-094937642ef8\" (UID: \"4bd61b3c-44d3-46c3-9e4e-094937642ef8\") " Dec 03 16:39:50 crc kubenswrapper[4998]: I1203 16:39:50.871431 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bd61b3c-44d3-46c3-9e4e-094937642ef8-inventory\") pod \"4bd61b3c-44d3-46c3-9e4e-094937642ef8\" (UID: \"4bd61b3c-44d3-46c3-9e4e-094937642ef8\") " Dec 03 16:39:50 crc kubenswrapper[4998]: I1203 16:39:50.876442 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bd61b3c-44d3-46c3-9e4e-094937642ef8-kube-api-access-7vkps" (OuterVolumeSpecName: "kube-api-access-7vkps") pod "4bd61b3c-44d3-46c3-9e4e-094937642ef8" (UID: "4bd61b3c-44d3-46c3-9e4e-094937642ef8"). InnerVolumeSpecName "kube-api-access-7vkps". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:39:50 crc kubenswrapper[4998]: I1203 16:39:50.899434 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bd61b3c-44d3-46c3-9e4e-094937642ef8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4bd61b3c-44d3-46c3-9e4e-094937642ef8" (UID: "4bd61b3c-44d3-46c3-9e4e-094937642ef8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:39:50 crc kubenswrapper[4998]: I1203 16:39:50.923389 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bd61b3c-44d3-46c3-9e4e-094937642ef8-inventory" (OuterVolumeSpecName: "inventory") pod "4bd61b3c-44d3-46c3-9e4e-094937642ef8" (UID: "4bd61b3c-44d3-46c3-9e4e-094937642ef8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:39:50 crc kubenswrapper[4998]: I1203 16:39:50.974919 4998 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bd61b3c-44d3-46c3-9e4e-094937642ef8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 16:39:50 crc kubenswrapper[4998]: I1203 16:39:50.975004 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vkps\" (UniqueName: \"kubernetes.io/projected/4bd61b3c-44d3-46c3-9e4e-094937642ef8-kube-api-access-7vkps\") on node \"crc\" DevicePath \"\"" Dec 03 16:39:50 crc kubenswrapper[4998]: I1203 16:39:50.975022 4998 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bd61b3c-44d3-46c3-9e4e-094937642ef8-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.397630 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v57vv" event={"ID":"4bd61b3c-44d3-46c3-9e4e-094937642ef8","Type":"ContainerDied","Data":"178b639a9aa4f14d1ae078dbc71029a0c503e5a977ab79ae9983eefb0dcaccc2"} Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.397670 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="178b639a9aa4f14d1ae078dbc71029a0c503e5a977ab79ae9983eefb0dcaccc2" Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.397683 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-v57vv" Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.506439 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-97zsr"] Dec 03 16:39:51 crc kubenswrapper[4998]: E1203 16:39:51.506986 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bd61b3c-44d3-46c3-9e4e-094937642ef8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.507009 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bd61b3c-44d3-46c3-9e4e-094937642ef8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 16:39:51 crc kubenswrapper[4998]: E1203 16:39:51.507037 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="145328e4-60b9-49c0-8e9a-36007a5dc2c3" containerName="extract-utilities" Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.507049 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="145328e4-60b9-49c0-8e9a-36007a5dc2c3" containerName="extract-utilities" Dec 03 16:39:51 crc kubenswrapper[4998]: E1203 16:39:51.507079 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="145328e4-60b9-49c0-8e9a-36007a5dc2c3" containerName="registry-server" Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.507087 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="145328e4-60b9-49c0-8e9a-36007a5dc2c3" containerName="registry-server" Dec 03 16:39:51 crc kubenswrapper[4998]: E1203 16:39:51.507104 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="145328e4-60b9-49c0-8e9a-36007a5dc2c3" containerName="extract-content" Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.507111 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="145328e4-60b9-49c0-8e9a-36007a5dc2c3" containerName="extract-content" Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.507358 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="145328e4-60b9-49c0-8e9a-36007a5dc2c3" containerName="registry-server" Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.507376 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bd61b3c-44d3-46c3-9e4e-094937642ef8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.508307 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-97zsr" Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.510888 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.511497 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.511831 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kcj9x" Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.515815 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.515866 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-97zsr"] Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.595780 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsj2g\" (UniqueName: \"kubernetes.io/projected/67c06662-6c0c-484f-b0b9-c61311305323-kube-api-access-wsj2g\") pod \"ssh-known-hosts-edpm-deployment-97zsr\" (UID: \"67c06662-6c0c-484f-b0b9-c61311305323\") " pod="openstack/ssh-known-hosts-edpm-deployment-97zsr" Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.596279 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/67c06662-6c0c-484f-b0b9-c61311305323-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-97zsr\" (UID: \"67c06662-6c0c-484f-b0b9-c61311305323\") " pod="openstack/ssh-known-hosts-edpm-deployment-97zsr" Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.596625 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/67c06662-6c0c-484f-b0b9-c61311305323-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-97zsr\" (UID: \"67c06662-6c0c-484f-b0b9-c61311305323\") " pod="openstack/ssh-known-hosts-edpm-deployment-97zsr" Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.698967 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsj2g\" (UniqueName: \"kubernetes.io/projected/67c06662-6c0c-484f-b0b9-c61311305323-kube-api-access-wsj2g\") pod \"ssh-known-hosts-edpm-deployment-97zsr\" (UID: \"67c06662-6c0c-484f-b0b9-c61311305323\") " pod="openstack/ssh-known-hosts-edpm-deployment-97zsr" Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.699147 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/67c06662-6c0c-484f-b0b9-c61311305323-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-97zsr\" (UID: \"67c06662-6c0c-484f-b0b9-c61311305323\") " pod="openstack/ssh-known-hosts-edpm-deployment-97zsr" Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.699557 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/67c06662-6c0c-484f-b0b9-c61311305323-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-97zsr\" (UID: \"67c06662-6c0c-484f-b0b9-c61311305323\") " pod="openstack/ssh-known-hosts-edpm-deployment-97zsr" Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.705146 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/67c06662-6c0c-484f-b0b9-c61311305323-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-97zsr\" (UID: \"67c06662-6c0c-484f-b0b9-c61311305323\") " pod="openstack/ssh-known-hosts-edpm-deployment-97zsr" Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.705729 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/67c06662-6c0c-484f-b0b9-c61311305323-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-97zsr\" (UID: \"67c06662-6c0c-484f-b0b9-c61311305323\") " pod="openstack/ssh-known-hosts-edpm-deployment-97zsr" Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.723423 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsj2g\" (UniqueName: \"kubernetes.io/projected/67c06662-6c0c-484f-b0b9-c61311305323-kube-api-access-wsj2g\") pod \"ssh-known-hosts-edpm-deployment-97zsr\" (UID: \"67c06662-6c0c-484f-b0b9-c61311305323\") " pod="openstack/ssh-known-hosts-edpm-deployment-97zsr" Dec 03 16:39:51 crc kubenswrapper[4998]: I1203 16:39:51.838009 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-97zsr" Dec 03 16:39:52 crc kubenswrapper[4998]: I1203 16:39:52.174154 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-97zsr"] Dec 03 16:39:52 crc kubenswrapper[4998]: I1203 16:39:52.434914 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-97zsr" event={"ID":"67c06662-6c0c-484f-b0b9-c61311305323","Type":"ContainerStarted","Data":"7a5b64acd2b1be28202557da6abffae5336fdad0a719169671a2340423df51cb"} Dec 03 16:39:53 crc kubenswrapper[4998]: I1203 16:39:53.443771 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-97zsr" event={"ID":"67c06662-6c0c-484f-b0b9-c61311305323","Type":"ContainerStarted","Data":"252345e74c018b4f13b80b90b049bb044794d776804d501ce7c0c5d7e2a119aa"} Dec 03 16:39:53 crc kubenswrapper[4998]: I1203 16:39:53.466174 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-97zsr" podStartSLOduration=1.956767629 podStartE2EDuration="2.466158361s" podCreationTimestamp="2025-12-03 16:39:51 +0000 UTC" firstStartedPulling="2025-12-03 16:39:52.181855548 +0000 UTC m=+2170.793555771" lastFinishedPulling="2025-12-03 16:39:52.69124628 +0000 UTC m=+2171.302946503" observedRunningTime="2025-12-03 16:39:53.462201351 +0000 UTC m=+2172.073901574" watchObservedRunningTime="2025-12-03 16:39:53.466158361 +0000 UTC m=+2172.077858584" Dec 03 16:39:57 crc kubenswrapper[4998]: I1203 16:39:57.111176 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:39:57 crc kubenswrapper[4998]: I1203 16:39:57.112389 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:39:57 crc kubenswrapper[4998]: I1203 16:39:57.112532 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:39:57 crc kubenswrapper[4998]: I1203 16:39:57.113483 4998 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155"} pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 16:39:57 crc kubenswrapper[4998]: I1203 16:39:57.113657 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" containerID="cri-o://8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" gracePeriod=600 Dec 03 16:39:57 crc kubenswrapper[4998]: E1203 16:39:57.243695 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:39:57 crc kubenswrapper[4998]: I1203 16:39:57.486222 4998 generic.go:334] "Generic (PLEG): container finished" podID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" exitCode=0 Dec 03 16:39:57 crc kubenswrapper[4998]: I1203 16:39:57.486299 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerDied","Data":"8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155"} Dec 03 16:39:57 crc kubenswrapper[4998]: I1203 16:39:57.486359 4998 scope.go:117] "RemoveContainer" containerID="9001df2bc0dc33071d80330946bfdea9e78811028533587a680aea91a458fee6" Dec 03 16:39:57 crc kubenswrapper[4998]: I1203 16:39:57.487031 4998 scope.go:117] "RemoveContainer" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" Dec 03 16:39:57 crc kubenswrapper[4998]: E1203 16:39:57.487422 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:40:00 crc kubenswrapper[4998]: I1203 16:40:00.537626 4998 generic.go:334] "Generic (PLEG): container finished" podID="67c06662-6c0c-484f-b0b9-c61311305323" containerID="252345e74c018b4f13b80b90b049bb044794d776804d501ce7c0c5d7e2a119aa" exitCode=0 Dec 03 16:40:00 crc kubenswrapper[4998]: I1203 16:40:00.537911 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-97zsr" event={"ID":"67c06662-6c0c-484f-b0b9-c61311305323","Type":"ContainerDied","Data":"252345e74c018b4f13b80b90b049bb044794d776804d501ce7c0c5d7e2a119aa"} Dec 03 16:40:01 crc kubenswrapper[4998]: I1203 16:40:01.998858 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-97zsr" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.048815 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsj2g\" (UniqueName: \"kubernetes.io/projected/67c06662-6c0c-484f-b0b9-c61311305323-kube-api-access-wsj2g\") pod \"67c06662-6c0c-484f-b0b9-c61311305323\" (UID: \"67c06662-6c0c-484f-b0b9-c61311305323\") " Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.048928 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/67c06662-6c0c-484f-b0b9-c61311305323-ssh-key-openstack-edpm-ipam\") pod \"67c06662-6c0c-484f-b0b9-c61311305323\" (UID: \"67c06662-6c0c-484f-b0b9-c61311305323\") " Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.048991 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/67c06662-6c0c-484f-b0b9-c61311305323-inventory-0\") pod \"67c06662-6c0c-484f-b0b9-c61311305323\" (UID: \"67c06662-6c0c-484f-b0b9-c61311305323\") " Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.055575 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67c06662-6c0c-484f-b0b9-c61311305323-kube-api-access-wsj2g" (OuterVolumeSpecName: "kube-api-access-wsj2g") pod "67c06662-6c0c-484f-b0b9-c61311305323" (UID: "67c06662-6c0c-484f-b0b9-c61311305323"). InnerVolumeSpecName "kube-api-access-wsj2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.091054 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67c06662-6c0c-484f-b0b9-c61311305323-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "67c06662-6c0c-484f-b0b9-c61311305323" (UID: "67c06662-6c0c-484f-b0b9-c61311305323"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.100211 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67c06662-6c0c-484f-b0b9-c61311305323-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "67c06662-6c0c-484f-b0b9-c61311305323" (UID: "67c06662-6c0c-484f-b0b9-c61311305323"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.152081 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsj2g\" (UniqueName: \"kubernetes.io/projected/67c06662-6c0c-484f-b0b9-c61311305323-kube-api-access-wsj2g\") on node \"crc\" DevicePath \"\"" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.152118 4998 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/67c06662-6c0c-484f-b0b9-c61311305323-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.152132 4998 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/67c06662-6c0c-484f-b0b9-c61311305323-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.557484 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-97zsr" event={"ID":"67c06662-6c0c-484f-b0b9-c61311305323","Type":"ContainerDied","Data":"7a5b64acd2b1be28202557da6abffae5336fdad0a719169671a2340423df51cb"} Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.557742 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a5b64acd2b1be28202557da6abffae5336fdad0a719169671a2340423df51cb" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.557526 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-97zsr" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.649885 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-29tst"] Dec 03 16:40:02 crc kubenswrapper[4998]: E1203 16:40:02.650419 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67c06662-6c0c-484f-b0b9-c61311305323" containerName="ssh-known-hosts-edpm-deployment" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.650437 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="67c06662-6c0c-484f-b0b9-c61311305323" containerName="ssh-known-hosts-edpm-deployment" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.650695 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="67c06662-6c0c-484f-b0b9-c61311305323" containerName="ssh-known-hosts-edpm-deployment" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.651386 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-29tst" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.655317 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.655357 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.655811 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.655880 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kcj9x" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.661145 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-29tst"] Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.767605 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c46460e9-7d52-435d-b5fd-2ebb81570dcf-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-29tst\" (UID: \"c46460e9-7d52-435d-b5fd-2ebb81570dcf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-29tst" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.767657 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c46460e9-7d52-435d-b5fd-2ebb81570dcf-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-29tst\" (UID: \"c46460e9-7d52-435d-b5fd-2ebb81570dcf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-29tst" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.767768 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6khxd\" (UniqueName: \"kubernetes.io/projected/c46460e9-7d52-435d-b5fd-2ebb81570dcf-kube-api-access-6khxd\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-29tst\" (UID: \"c46460e9-7d52-435d-b5fd-2ebb81570dcf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-29tst" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.868874 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6khxd\" (UniqueName: \"kubernetes.io/projected/c46460e9-7d52-435d-b5fd-2ebb81570dcf-kube-api-access-6khxd\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-29tst\" (UID: \"c46460e9-7d52-435d-b5fd-2ebb81570dcf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-29tst" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.869431 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c46460e9-7d52-435d-b5fd-2ebb81570dcf-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-29tst\" (UID: \"c46460e9-7d52-435d-b5fd-2ebb81570dcf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-29tst" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.869491 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c46460e9-7d52-435d-b5fd-2ebb81570dcf-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-29tst\" (UID: \"c46460e9-7d52-435d-b5fd-2ebb81570dcf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-29tst" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.875002 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c46460e9-7d52-435d-b5fd-2ebb81570dcf-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-29tst\" (UID: \"c46460e9-7d52-435d-b5fd-2ebb81570dcf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-29tst" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.878115 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c46460e9-7d52-435d-b5fd-2ebb81570dcf-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-29tst\" (UID: \"c46460e9-7d52-435d-b5fd-2ebb81570dcf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-29tst" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.893898 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6khxd\" (UniqueName: \"kubernetes.io/projected/c46460e9-7d52-435d-b5fd-2ebb81570dcf-kube-api-access-6khxd\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-29tst\" (UID: \"c46460e9-7d52-435d-b5fd-2ebb81570dcf\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-29tst" Dec 03 16:40:02 crc kubenswrapper[4998]: I1203 16:40:02.968595 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-29tst" Dec 03 16:40:03 crc kubenswrapper[4998]: I1203 16:40:03.515077 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-29tst"] Dec 03 16:40:03 crc kubenswrapper[4998]: I1203 16:40:03.575723 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-29tst" event={"ID":"c46460e9-7d52-435d-b5fd-2ebb81570dcf","Type":"ContainerStarted","Data":"ce3d9673bb7b009a61a13e2a92011a732ec53cca4dd8f6983e5cbc099f327148"} Dec 03 16:40:04 crc kubenswrapper[4998]: I1203 16:40:04.593718 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-29tst" event={"ID":"c46460e9-7d52-435d-b5fd-2ebb81570dcf","Type":"ContainerStarted","Data":"1ddbc248c245bf7077824209c4e96dc8ea46ef3d774b4596f00410b1bba4fdf9"} Dec 03 16:40:04 crc kubenswrapper[4998]: I1203 16:40:04.627099 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-29tst" podStartSLOduration=2.240778087 podStartE2EDuration="2.627080713s" podCreationTimestamp="2025-12-03 16:40:02 +0000 UTC" firstStartedPulling="2025-12-03 16:40:03.51748503 +0000 UTC m=+2182.129185293" lastFinishedPulling="2025-12-03 16:40:03.903787696 +0000 UTC m=+2182.515487919" observedRunningTime="2025-12-03 16:40:04.61341263 +0000 UTC m=+2183.225112863" watchObservedRunningTime="2025-12-03 16:40:04.627080713 +0000 UTC m=+2183.238780936" Dec 03 16:40:05 crc kubenswrapper[4998]: I1203 16:40:05.691300 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jf8mb"] Dec 03 16:40:05 crc kubenswrapper[4998]: I1203 16:40:05.699088 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jf8mb"] Dec 03 16:40:05 crc kubenswrapper[4998]: I1203 16:40:05.699228 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jf8mb" Dec 03 16:40:05 crc kubenswrapper[4998]: I1203 16:40:05.728878 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c599ccba-c9ef-4044-bfc9-fcf92867b12b-catalog-content\") pod \"community-operators-jf8mb\" (UID: \"c599ccba-c9ef-4044-bfc9-fcf92867b12b\") " pod="openshift-marketplace/community-operators-jf8mb" Dec 03 16:40:05 crc kubenswrapper[4998]: I1203 16:40:05.729142 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c599ccba-c9ef-4044-bfc9-fcf92867b12b-utilities\") pod \"community-operators-jf8mb\" (UID: \"c599ccba-c9ef-4044-bfc9-fcf92867b12b\") " pod="openshift-marketplace/community-operators-jf8mb" Dec 03 16:40:05 crc kubenswrapper[4998]: I1203 16:40:05.729270 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwpzz\" (UniqueName: \"kubernetes.io/projected/c599ccba-c9ef-4044-bfc9-fcf92867b12b-kube-api-access-vwpzz\") pod \"community-operators-jf8mb\" (UID: \"c599ccba-c9ef-4044-bfc9-fcf92867b12b\") " pod="openshift-marketplace/community-operators-jf8mb" Dec 03 16:40:05 crc kubenswrapper[4998]: I1203 16:40:05.832309 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c599ccba-c9ef-4044-bfc9-fcf92867b12b-catalog-content\") pod \"community-operators-jf8mb\" (UID: \"c599ccba-c9ef-4044-bfc9-fcf92867b12b\") " pod="openshift-marketplace/community-operators-jf8mb" Dec 03 16:40:05 crc kubenswrapper[4998]: I1203 16:40:05.832364 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c599ccba-c9ef-4044-bfc9-fcf92867b12b-utilities\") pod \"community-operators-jf8mb\" (UID: \"c599ccba-c9ef-4044-bfc9-fcf92867b12b\") " pod="openshift-marketplace/community-operators-jf8mb" Dec 03 16:40:05 crc kubenswrapper[4998]: I1203 16:40:05.832420 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwpzz\" (UniqueName: \"kubernetes.io/projected/c599ccba-c9ef-4044-bfc9-fcf92867b12b-kube-api-access-vwpzz\") pod \"community-operators-jf8mb\" (UID: \"c599ccba-c9ef-4044-bfc9-fcf92867b12b\") " pod="openshift-marketplace/community-operators-jf8mb" Dec 03 16:40:05 crc kubenswrapper[4998]: I1203 16:40:05.833202 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c599ccba-c9ef-4044-bfc9-fcf92867b12b-catalog-content\") pod \"community-operators-jf8mb\" (UID: \"c599ccba-c9ef-4044-bfc9-fcf92867b12b\") " pod="openshift-marketplace/community-operators-jf8mb" Dec 03 16:40:05 crc kubenswrapper[4998]: I1203 16:40:05.833459 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c599ccba-c9ef-4044-bfc9-fcf92867b12b-utilities\") pod \"community-operators-jf8mb\" (UID: \"c599ccba-c9ef-4044-bfc9-fcf92867b12b\") " pod="openshift-marketplace/community-operators-jf8mb" Dec 03 16:40:05 crc kubenswrapper[4998]: I1203 16:40:05.850806 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwpzz\" (UniqueName: \"kubernetes.io/projected/c599ccba-c9ef-4044-bfc9-fcf92867b12b-kube-api-access-vwpzz\") pod \"community-operators-jf8mb\" (UID: \"c599ccba-c9ef-4044-bfc9-fcf92867b12b\") " pod="openshift-marketplace/community-operators-jf8mb" Dec 03 16:40:06 crc kubenswrapper[4998]: I1203 16:40:06.024452 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jf8mb" Dec 03 16:40:06 crc kubenswrapper[4998]: I1203 16:40:06.590507 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jf8mb"] Dec 03 16:40:06 crc kubenswrapper[4998]: I1203 16:40:06.612988 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jf8mb" event={"ID":"c599ccba-c9ef-4044-bfc9-fcf92867b12b","Type":"ContainerStarted","Data":"830673e708e3d21c56fe4c15a9a3e9c710009fb0f99c32f7a1358fae2e421760"} Dec 03 16:40:07 crc kubenswrapper[4998]: I1203 16:40:07.645476 4998 generic.go:334] "Generic (PLEG): container finished" podID="c599ccba-c9ef-4044-bfc9-fcf92867b12b" containerID="d3e6b10baf75722b5b2a8ffd61f113961395d9e042e4da0180588f6dc9bed333" exitCode=0 Dec 03 16:40:07 crc kubenswrapper[4998]: I1203 16:40:07.645585 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jf8mb" event={"ID":"c599ccba-c9ef-4044-bfc9-fcf92867b12b","Type":"ContainerDied","Data":"d3e6b10baf75722b5b2a8ffd61f113961395d9e042e4da0180588f6dc9bed333"} Dec 03 16:40:09 crc kubenswrapper[4998]: I1203 16:40:09.665032 4998 generic.go:334] "Generic (PLEG): container finished" podID="c599ccba-c9ef-4044-bfc9-fcf92867b12b" containerID="f55d62b6389c3c75d50b1214e2984526046587dedd57bd3153452c6e7cbb8e24" exitCode=0 Dec 03 16:40:09 crc kubenswrapper[4998]: I1203 16:40:09.665122 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jf8mb" event={"ID":"c599ccba-c9ef-4044-bfc9-fcf92867b12b","Type":"ContainerDied","Data":"f55d62b6389c3c75d50b1214e2984526046587dedd57bd3153452c6e7cbb8e24"} Dec 03 16:40:10 crc kubenswrapper[4998]: I1203 16:40:10.675843 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jf8mb" event={"ID":"c599ccba-c9ef-4044-bfc9-fcf92867b12b","Type":"ContainerStarted","Data":"137523470c6e7f7a0c3ebb23ecc9caf72a896a0c71a090db91514ae66376c715"} Dec 03 16:40:10 crc kubenswrapper[4998]: I1203 16:40:10.695174 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jf8mb" podStartSLOduration=3.271676088 podStartE2EDuration="5.695157146s" podCreationTimestamp="2025-12-03 16:40:05 +0000 UTC" firstStartedPulling="2025-12-03 16:40:07.648049631 +0000 UTC m=+2186.259749854" lastFinishedPulling="2025-12-03 16:40:10.071530689 +0000 UTC m=+2188.683230912" observedRunningTime="2025-12-03 16:40:10.692373376 +0000 UTC m=+2189.304073599" watchObservedRunningTime="2025-12-03 16:40:10.695157146 +0000 UTC m=+2189.306857359" Dec 03 16:40:12 crc kubenswrapper[4998]: I1203 16:40:12.679251 4998 scope.go:117] "RemoveContainer" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" Dec 03 16:40:12 crc kubenswrapper[4998]: E1203 16:40:12.679901 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:40:12 crc kubenswrapper[4998]: I1203 16:40:12.694643 4998 generic.go:334] "Generic (PLEG): container finished" podID="c46460e9-7d52-435d-b5fd-2ebb81570dcf" containerID="1ddbc248c245bf7077824209c4e96dc8ea46ef3d774b4596f00410b1bba4fdf9" exitCode=0 Dec 03 16:40:12 crc kubenswrapper[4998]: I1203 16:40:12.694688 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-29tst" event={"ID":"c46460e9-7d52-435d-b5fd-2ebb81570dcf","Type":"ContainerDied","Data":"1ddbc248c245bf7077824209c4e96dc8ea46ef3d774b4596f00410b1bba4fdf9"} Dec 03 16:40:14 crc kubenswrapper[4998]: I1203 16:40:14.148645 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-29tst" Dec 03 16:40:14 crc kubenswrapper[4998]: I1203 16:40:14.197947 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c46460e9-7d52-435d-b5fd-2ebb81570dcf-inventory\") pod \"c46460e9-7d52-435d-b5fd-2ebb81570dcf\" (UID: \"c46460e9-7d52-435d-b5fd-2ebb81570dcf\") " Dec 03 16:40:14 crc kubenswrapper[4998]: I1203 16:40:14.198264 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c46460e9-7d52-435d-b5fd-2ebb81570dcf-ssh-key\") pod \"c46460e9-7d52-435d-b5fd-2ebb81570dcf\" (UID: \"c46460e9-7d52-435d-b5fd-2ebb81570dcf\") " Dec 03 16:40:14 crc kubenswrapper[4998]: I1203 16:40:14.198302 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6khxd\" (UniqueName: \"kubernetes.io/projected/c46460e9-7d52-435d-b5fd-2ebb81570dcf-kube-api-access-6khxd\") pod \"c46460e9-7d52-435d-b5fd-2ebb81570dcf\" (UID: \"c46460e9-7d52-435d-b5fd-2ebb81570dcf\") " Dec 03 16:40:14 crc kubenswrapper[4998]: I1203 16:40:14.204074 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c46460e9-7d52-435d-b5fd-2ebb81570dcf-kube-api-access-6khxd" (OuterVolumeSpecName: "kube-api-access-6khxd") pod "c46460e9-7d52-435d-b5fd-2ebb81570dcf" (UID: "c46460e9-7d52-435d-b5fd-2ebb81570dcf"). InnerVolumeSpecName "kube-api-access-6khxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:40:14 crc kubenswrapper[4998]: I1203 16:40:14.230615 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c46460e9-7d52-435d-b5fd-2ebb81570dcf-inventory" (OuterVolumeSpecName: "inventory") pod "c46460e9-7d52-435d-b5fd-2ebb81570dcf" (UID: "c46460e9-7d52-435d-b5fd-2ebb81570dcf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:40:14 crc kubenswrapper[4998]: I1203 16:40:14.242276 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c46460e9-7d52-435d-b5fd-2ebb81570dcf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c46460e9-7d52-435d-b5fd-2ebb81570dcf" (UID: "c46460e9-7d52-435d-b5fd-2ebb81570dcf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:40:14 crc kubenswrapper[4998]: I1203 16:40:14.300460 4998 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c46460e9-7d52-435d-b5fd-2ebb81570dcf-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 16:40:14 crc kubenswrapper[4998]: I1203 16:40:14.300705 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6khxd\" (UniqueName: \"kubernetes.io/projected/c46460e9-7d52-435d-b5fd-2ebb81570dcf-kube-api-access-6khxd\") on node \"crc\" DevicePath \"\"" Dec 03 16:40:14 crc kubenswrapper[4998]: I1203 16:40:14.300861 4998 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c46460e9-7d52-435d-b5fd-2ebb81570dcf-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 16:40:14 crc kubenswrapper[4998]: I1203 16:40:14.714156 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-29tst" event={"ID":"c46460e9-7d52-435d-b5fd-2ebb81570dcf","Type":"ContainerDied","Data":"ce3d9673bb7b009a61a13e2a92011a732ec53cca4dd8f6983e5cbc099f327148"} Dec 03 16:40:14 crc kubenswrapper[4998]: I1203 16:40:14.714185 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-29tst" Dec 03 16:40:14 crc kubenswrapper[4998]: I1203 16:40:14.714192 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce3d9673bb7b009a61a13e2a92011a732ec53cca4dd8f6983e5cbc099f327148" Dec 03 16:40:14 crc kubenswrapper[4998]: I1203 16:40:14.820926 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s"] Dec 03 16:40:14 crc kubenswrapper[4998]: E1203 16:40:14.821389 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c46460e9-7d52-435d-b5fd-2ebb81570dcf" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 16:40:14 crc kubenswrapper[4998]: I1203 16:40:14.821409 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="c46460e9-7d52-435d-b5fd-2ebb81570dcf" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 16:40:14 crc kubenswrapper[4998]: I1203 16:40:14.821679 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="c46460e9-7d52-435d-b5fd-2ebb81570dcf" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 16:40:14 crc kubenswrapper[4998]: I1203 16:40:14.822505 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s" Dec 03 16:40:14 crc kubenswrapper[4998]: I1203 16:40:14.825573 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 16:40:14 crc kubenswrapper[4998]: I1203 16:40:14.825861 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 16:40:14 crc kubenswrapper[4998]: I1203 16:40:14.827136 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kcj9x" Dec 03 16:40:14 crc kubenswrapper[4998]: I1203 16:40:14.833435 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 16:40:14 crc kubenswrapper[4998]: I1203 16:40:14.842837 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s"] Dec 03 16:40:14 crc kubenswrapper[4998]: I1203 16:40:14.913904 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5c35764-7c7b-4346-b54b-6265ec6bca06-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s\" (UID: \"d5c35764-7c7b-4346-b54b-6265ec6bca06\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s" Dec 03 16:40:14 crc kubenswrapper[4998]: I1203 16:40:14.914275 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5c35764-7c7b-4346-b54b-6265ec6bca06-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s\" (UID: \"d5c35764-7c7b-4346-b54b-6265ec6bca06\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s" Dec 03 16:40:14 crc kubenswrapper[4998]: I1203 16:40:14.914309 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6crr\" (UniqueName: \"kubernetes.io/projected/d5c35764-7c7b-4346-b54b-6265ec6bca06-kube-api-access-p6crr\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s\" (UID: \"d5c35764-7c7b-4346-b54b-6265ec6bca06\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s" Dec 03 16:40:15 crc kubenswrapper[4998]: I1203 16:40:15.015223 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5c35764-7c7b-4346-b54b-6265ec6bca06-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s\" (UID: \"d5c35764-7c7b-4346-b54b-6265ec6bca06\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s" Dec 03 16:40:15 crc kubenswrapper[4998]: I1203 16:40:15.015281 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6crr\" (UniqueName: \"kubernetes.io/projected/d5c35764-7c7b-4346-b54b-6265ec6bca06-kube-api-access-p6crr\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s\" (UID: \"d5c35764-7c7b-4346-b54b-6265ec6bca06\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s" Dec 03 16:40:15 crc kubenswrapper[4998]: I1203 16:40:15.015428 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5c35764-7c7b-4346-b54b-6265ec6bca06-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s\" (UID: \"d5c35764-7c7b-4346-b54b-6265ec6bca06\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s" Dec 03 16:40:15 crc kubenswrapper[4998]: I1203 16:40:15.023792 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5c35764-7c7b-4346-b54b-6265ec6bca06-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s\" (UID: \"d5c35764-7c7b-4346-b54b-6265ec6bca06\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s" Dec 03 16:40:15 crc kubenswrapper[4998]: I1203 16:40:15.023850 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5c35764-7c7b-4346-b54b-6265ec6bca06-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s\" (UID: \"d5c35764-7c7b-4346-b54b-6265ec6bca06\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s" Dec 03 16:40:15 crc kubenswrapper[4998]: I1203 16:40:15.041072 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6crr\" (UniqueName: \"kubernetes.io/projected/d5c35764-7c7b-4346-b54b-6265ec6bca06-kube-api-access-p6crr\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s\" (UID: \"d5c35764-7c7b-4346-b54b-6265ec6bca06\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s" Dec 03 16:40:15 crc kubenswrapper[4998]: I1203 16:40:15.141694 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s" Dec 03 16:40:15 crc kubenswrapper[4998]: I1203 16:40:15.721381 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s"] Dec 03 16:40:16 crc kubenswrapper[4998]: I1203 16:40:16.025577 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jf8mb" Dec 03 16:40:16 crc kubenswrapper[4998]: I1203 16:40:16.025952 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jf8mb" Dec 03 16:40:16 crc kubenswrapper[4998]: I1203 16:40:16.081505 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jf8mb" Dec 03 16:40:16 crc kubenswrapper[4998]: I1203 16:40:16.732495 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s" event={"ID":"d5c35764-7c7b-4346-b54b-6265ec6bca06","Type":"ContainerStarted","Data":"c918c2f509dbc83240aab1294a014d82259b38537fd71a9ed7f1e4ec5d266b67"} Dec 03 16:40:16 crc kubenswrapper[4998]: I1203 16:40:16.732893 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s" event={"ID":"d5c35764-7c7b-4346-b54b-6265ec6bca06","Type":"ContainerStarted","Data":"c67e70e678a08ac8603833d9464f523abe6414720120fca1cb6bbdf889d29922"} Dec 03 16:40:16 crc kubenswrapper[4998]: I1203 16:40:16.753905 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s" podStartSLOduration=2.157061819 podStartE2EDuration="2.753887964s" podCreationTimestamp="2025-12-03 16:40:14 +0000 UTC" firstStartedPulling="2025-12-03 16:40:15.723467807 +0000 UTC m=+2194.335168030" lastFinishedPulling="2025-12-03 16:40:16.320293912 +0000 UTC m=+2194.931994175" observedRunningTime="2025-12-03 16:40:16.749945205 +0000 UTC m=+2195.361645438" watchObservedRunningTime="2025-12-03 16:40:16.753887964 +0000 UTC m=+2195.365588197" Dec 03 16:40:16 crc kubenswrapper[4998]: I1203 16:40:16.823599 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jf8mb" Dec 03 16:40:16 crc kubenswrapper[4998]: I1203 16:40:16.889397 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jf8mb"] Dec 03 16:40:18 crc kubenswrapper[4998]: I1203 16:40:18.749403 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jf8mb" podUID="c599ccba-c9ef-4044-bfc9-fcf92867b12b" containerName="registry-server" containerID="cri-o://137523470c6e7f7a0c3ebb23ecc9caf72a896a0c71a090db91514ae66376c715" gracePeriod=2 Dec 03 16:40:19 crc kubenswrapper[4998]: I1203 16:40:19.196828 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jf8mb" Dec 03 16:40:19 crc kubenswrapper[4998]: I1203 16:40:19.302337 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c599ccba-c9ef-4044-bfc9-fcf92867b12b-utilities\") pod \"c599ccba-c9ef-4044-bfc9-fcf92867b12b\" (UID: \"c599ccba-c9ef-4044-bfc9-fcf92867b12b\") " Dec 03 16:40:19 crc kubenswrapper[4998]: I1203 16:40:19.302472 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c599ccba-c9ef-4044-bfc9-fcf92867b12b-catalog-content\") pod \"c599ccba-c9ef-4044-bfc9-fcf92867b12b\" (UID: \"c599ccba-c9ef-4044-bfc9-fcf92867b12b\") " Dec 03 16:40:19 crc kubenswrapper[4998]: I1203 16:40:19.302587 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwpzz\" (UniqueName: \"kubernetes.io/projected/c599ccba-c9ef-4044-bfc9-fcf92867b12b-kube-api-access-vwpzz\") pod \"c599ccba-c9ef-4044-bfc9-fcf92867b12b\" (UID: \"c599ccba-c9ef-4044-bfc9-fcf92867b12b\") " Dec 03 16:40:19 crc kubenswrapper[4998]: I1203 16:40:19.304648 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c599ccba-c9ef-4044-bfc9-fcf92867b12b-utilities" (OuterVolumeSpecName: "utilities") pod "c599ccba-c9ef-4044-bfc9-fcf92867b12b" (UID: "c599ccba-c9ef-4044-bfc9-fcf92867b12b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:40:19 crc kubenswrapper[4998]: I1203 16:40:19.315927 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c599ccba-c9ef-4044-bfc9-fcf92867b12b-kube-api-access-vwpzz" (OuterVolumeSpecName: "kube-api-access-vwpzz") pod "c599ccba-c9ef-4044-bfc9-fcf92867b12b" (UID: "c599ccba-c9ef-4044-bfc9-fcf92867b12b"). InnerVolumeSpecName "kube-api-access-vwpzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:40:19 crc kubenswrapper[4998]: I1203 16:40:19.409495 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwpzz\" (UniqueName: \"kubernetes.io/projected/c599ccba-c9ef-4044-bfc9-fcf92867b12b-kube-api-access-vwpzz\") on node \"crc\" DevicePath \"\"" Dec 03 16:40:19 crc kubenswrapper[4998]: I1203 16:40:19.409533 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c599ccba-c9ef-4044-bfc9-fcf92867b12b-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:40:19 crc kubenswrapper[4998]: I1203 16:40:19.763031 4998 generic.go:334] "Generic (PLEG): container finished" podID="c599ccba-c9ef-4044-bfc9-fcf92867b12b" containerID="137523470c6e7f7a0c3ebb23ecc9caf72a896a0c71a090db91514ae66376c715" exitCode=0 Dec 03 16:40:19 crc kubenswrapper[4998]: I1203 16:40:19.763084 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jf8mb" event={"ID":"c599ccba-c9ef-4044-bfc9-fcf92867b12b","Type":"ContainerDied","Data":"137523470c6e7f7a0c3ebb23ecc9caf72a896a0c71a090db91514ae66376c715"} Dec 03 16:40:19 crc kubenswrapper[4998]: I1203 16:40:19.763109 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jf8mb" Dec 03 16:40:19 crc kubenswrapper[4998]: I1203 16:40:19.763133 4998 scope.go:117] "RemoveContainer" containerID="137523470c6e7f7a0c3ebb23ecc9caf72a896a0c71a090db91514ae66376c715" Dec 03 16:40:19 crc kubenswrapper[4998]: I1203 16:40:19.763118 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jf8mb" event={"ID":"c599ccba-c9ef-4044-bfc9-fcf92867b12b","Type":"ContainerDied","Data":"830673e708e3d21c56fe4c15a9a3e9c710009fb0f99c32f7a1358fae2e421760"} Dec 03 16:40:19 crc kubenswrapper[4998]: I1203 16:40:19.783428 4998 scope.go:117] "RemoveContainer" containerID="f55d62b6389c3c75d50b1214e2984526046587dedd57bd3153452c6e7cbb8e24" Dec 03 16:40:19 crc kubenswrapper[4998]: I1203 16:40:19.828309 4998 scope.go:117] "RemoveContainer" containerID="d3e6b10baf75722b5b2a8ffd61f113961395d9e042e4da0180588f6dc9bed333" Dec 03 16:40:19 crc kubenswrapper[4998]: I1203 16:40:19.856348 4998 scope.go:117] "RemoveContainer" containerID="137523470c6e7f7a0c3ebb23ecc9caf72a896a0c71a090db91514ae66376c715" Dec 03 16:40:19 crc kubenswrapper[4998]: E1203 16:40:19.856874 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"137523470c6e7f7a0c3ebb23ecc9caf72a896a0c71a090db91514ae66376c715\": container with ID starting with 137523470c6e7f7a0c3ebb23ecc9caf72a896a0c71a090db91514ae66376c715 not found: ID does not exist" containerID="137523470c6e7f7a0c3ebb23ecc9caf72a896a0c71a090db91514ae66376c715" Dec 03 16:40:19 crc kubenswrapper[4998]: I1203 16:40:19.856915 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"137523470c6e7f7a0c3ebb23ecc9caf72a896a0c71a090db91514ae66376c715"} err="failed to get container status \"137523470c6e7f7a0c3ebb23ecc9caf72a896a0c71a090db91514ae66376c715\": rpc error: code = NotFound desc = could not find container \"137523470c6e7f7a0c3ebb23ecc9caf72a896a0c71a090db91514ae66376c715\": container with ID starting with 137523470c6e7f7a0c3ebb23ecc9caf72a896a0c71a090db91514ae66376c715 not found: ID does not exist" Dec 03 16:40:19 crc kubenswrapper[4998]: I1203 16:40:19.856941 4998 scope.go:117] "RemoveContainer" containerID="f55d62b6389c3c75d50b1214e2984526046587dedd57bd3153452c6e7cbb8e24" Dec 03 16:40:19 crc kubenswrapper[4998]: E1203 16:40:19.857515 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f55d62b6389c3c75d50b1214e2984526046587dedd57bd3153452c6e7cbb8e24\": container with ID starting with f55d62b6389c3c75d50b1214e2984526046587dedd57bd3153452c6e7cbb8e24 not found: ID does not exist" containerID="f55d62b6389c3c75d50b1214e2984526046587dedd57bd3153452c6e7cbb8e24" Dec 03 16:40:19 crc kubenswrapper[4998]: I1203 16:40:19.857629 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f55d62b6389c3c75d50b1214e2984526046587dedd57bd3153452c6e7cbb8e24"} err="failed to get container status \"f55d62b6389c3c75d50b1214e2984526046587dedd57bd3153452c6e7cbb8e24\": rpc error: code = NotFound desc = could not find container \"f55d62b6389c3c75d50b1214e2984526046587dedd57bd3153452c6e7cbb8e24\": container with ID starting with f55d62b6389c3c75d50b1214e2984526046587dedd57bd3153452c6e7cbb8e24 not found: ID does not exist" Dec 03 16:40:19 crc kubenswrapper[4998]: I1203 16:40:19.857729 4998 scope.go:117] "RemoveContainer" containerID="d3e6b10baf75722b5b2a8ffd61f113961395d9e042e4da0180588f6dc9bed333" Dec 03 16:40:19 crc kubenswrapper[4998]: E1203 16:40:19.858169 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3e6b10baf75722b5b2a8ffd61f113961395d9e042e4da0180588f6dc9bed333\": container with ID starting with d3e6b10baf75722b5b2a8ffd61f113961395d9e042e4da0180588f6dc9bed333 not found: ID does not exist" containerID="d3e6b10baf75722b5b2a8ffd61f113961395d9e042e4da0180588f6dc9bed333" Dec 03 16:40:19 crc kubenswrapper[4998]: I1203 16:40:19.858200 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3e6b10baf75722b5b2a8ffd61f113961395d9e042e4da0180588f6dc9bed333"} err="failed to get container status \"d3e6b10baf75722b5b2a8ffd61f113961395d9e042e4da0180588f6dc9bed333\": rpc error: code = NotFound desc = could not find container \"d3e6b10baf75722b5b2a8ffd61f113961395d9e042e4da0180588f6dc9bed333\": container with ID starting with d3e6b10baf75722b5b2a8ffd61f113961395d9e042e4da0180588f6dc9bed333 not found: ID does not exist" Dec 03 16:40:19 crc kubenswrapper[4998]: I1203 16:40:19.859215 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c599ccba-c9ef-4044-bfc9-fcf92867b12b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c599ccba-c9ef-4044-bfc9-fcf92867b12b" (UID: "c599ccba-c9ef-4044-bfc9-fcf92867b12b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:40:19 crc kubenswrapper[4998]: I1203 16:40:19.920309 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c599ccba-c9ef-4044-bfc9-fcf92867b12b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:40:20 crc kubenswrapper[4998]: I1203 16:40:20.107016 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jf8mb"] Dec 03 16:40:20 crc kubenswrapper[4998]: I1203 16:40:20.120168 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jf8mb"] Dec 03 16:40:21 crc kubenswrapper[4998]: I1203 16:40:21.701345 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c599ccba-c9ef-4044-bfc9-fcf92867b12b" path="/var/lib/kubelet/pods/c599ccba-c9ef-4044-bfc9-fcf92867b12b/volumes" Dec 03 16:40:23 crc kubenswrapper[4998]: I1203 16:40:23.677771 4998 scope.go:117] "RemoveContainer" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" Dec 03 16:40:23 crc kubenswrapper[4998]: E1203 16:40:23.678020 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:40:26 crc kubenswrapper[4998]: I1203 16:40:26.843953 4998 generic.go:334] "Generic (PLEG): container finished" podID="d5c35764-7c7b-4346-b54b-6265ec6bca06" containerID="c918c2f509dbc83240aab1294a014d82259b38537fd71a9ed7f1e4ec5d266b67" exitCode=0 Dec 03 16:40:26 crc kubenswrapper[4998]: I1203 16:40:26.844147 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s" event={"ID":"d5c35764-7c7b-4346-b54b-6265ec6bca06","Type":"ContainerDied","Data":"c918c2f509dbc83240aab1294a014d82259b38537fd71a9ed7f1e4ec5d266b67"} Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.291252 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s" Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.485397 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6crr\" (UniqueName: \"kubernetes.io/projected/d5c35764-7c7b-4346-b54b-6265ec6bca06-kube-api-access-p6crr\") pod \"d5c35764-7c7b-4346-b54b-6265ec6bca06\" (UID: \"d5c35764-7c7b-4346-b54b-6265ec6bca06\") " Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.485511 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5c35764-7c7b-4346-b54b-6265ec6bca06-ssh-key\") pod \"d5c35764-7c7b-4346-b54b-6265ec6bca06\" (UID: \"d5c35764-7c7b-4346-b54b-6265ec6bca06\") " Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.485552 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5c35764-7c7b-4346-b54b-6265ec6bca06-inventory\") pod \"d5c35764-7c7b-4346-b54b-6265ec6bca06\" (UID: \"d5c35764-7c7b-4346-b54b-6265ec6bca06\") " Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.501125 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5c35764-7c7b-4346-b54b-6265ec6bca06-kube-api-access-p6crr" (OuterVolumeSpecName: "kube-api-access-p6crr") pod "d5c35764-7c7b-4346-b54b-6265ec6bca06" (UID: "d5c35764-7c7b-4346-b54b-6265ec6bca06"). InnerVolumeSpecName "kube-api-access-p6crr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.555908 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5c35764-7c7b-4346-b54b-6265ec6bca06-inventory" (OuterVolumeSpecName: "inventory") pod "d5c35764-7c7b-4346-b54b-6265ec6bca06" (UID: "d5c35764-7c7b-4346-b54b-6265ec6bca06"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.577954 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5c35764-7c7b-4346-b54b-6265ec6bca06-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d5c35764-7c7b-4346-b54b-6265ec6bca06" (UID: "d5c35764-7c7b-4346-b54b-6265ec6bca06"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.588084 4998 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5c35764-7c7b-4346-b54b-6265ec6bca06-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.588390 4998 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d5c35764-7c7b-4346-b54b-6265ec6bca06-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.588402 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6crr\" (UniqueName: \"kubernetes.io/projected/d5c35764-7c7b-4346-b54b-6265ec6bca06-kube-api-access-p6crr\") on node \"crc\" DevicePath \"\"" Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.862384 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s" event={"ID":"d5c35764-7c7b-4346-b54b-6265ec6bca06","Type":"ContainerDied","Data":"c67e70e678a08ac8603833d9464f523abe6414720120fca1cb6bbdf889d29922"} Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.862420 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c67e70e678a08ac8603833d9464f523abe6414720120fca1cb6bbdf889d29922" Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.862445 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s" Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.948717 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf"] Dec 03 16:40:28 crc kubenswrapper[4998]: E1203 16:40:28.949932 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c599ccba-c9ef-4044-bfc9-fcf92867b12b" containerName="registry-server" Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.950056 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="c599ccba-c9ef-4044-bfc9-fcf92867b12b" containerName="registry-server" Dec 03 16:40:28 crc kubenswrapper[4998]: E1203 16:40:28.950171 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5c35764-7c7b-4346-b54b-6265ec6bca06" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.950246 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5c35764-7c7b-4346-b54b-6265ec6bca06" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 16:40:28 crc kubenswrapper[4998]: E1203 16:40:28.950334 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c599ccba-c9ef-4044-bfc9-fcf92867b12b" containerName="extract-utilities" Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.950406 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="c599ccba-c9ef-4044-bfc9-fcf92867b12b" containerName="extract-utilities" Dec 03 16:40:28 crc kubenswrapper[4998]: E1203 16:40:28.950541 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c599ccba-c9ef-4044-bfc9-fcf92867b12b" containerName="extract-content" Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.950642 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="c599ccba-c9ef-4044-bfc9-fcf92867b12b" containerName="extract-content" Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.950994 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="c599ccba-c9ef-4044-bfc9-fcf92867b12b" containerName="registry-server" Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.951139 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5c35764-7c7b-4346-b54b-6265ec6bca06" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.952064 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.956072 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.956312 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.956438 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.956588 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.956851 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.957400 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kcj9x" Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.957570 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.959563 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 16:40:28 crc kubenswrapper[4998]: I1203 16:40:28.970676 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf"] Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.101867 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.101923 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.101984 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.102053 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.102081 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.102201 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.102278 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.102344 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.102378 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.102447 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.102490 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.102611 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.102721 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.102806 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thzwb\" (UniqueName: \"kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-kube-api-access-thzwb\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.205657 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.205781 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.205864 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.206666 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.206736 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.207162 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.207287 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thzwb\" (UniqueName: \"kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-kube-api-access-thzwb\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.207332 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.207350 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.207381 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.207411 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.207432 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.207488 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.207525 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.220563 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.221339 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.221542 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.221782 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.222518 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.223258 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.224073 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.224299 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.224573 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thzwb\" (UniqueName: \"kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-kube-api-access-thzwb\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.224673 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.225145 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.226518 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.226601 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.234534 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.280324 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.807467 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf"] Dec 03 16:40:29 crc kubenswrapper[4998]: I1203 16:40:29.873169 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" event={"ID":"c5cebbf5-5350-42c2-8908-8a508995a72c","Type":"ContainerStarted","Data":"bc206d8364647be875a9f54bd835d6204247b03a1fdb1e28236e77c03d6b4ede"} Dec 03 16:40:30 crc kubenswrapper[4998]: I1203 16:40:30.883212 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" event={"ID":"c5cebbf5-5350-42c2-8908-8a508995a72c","Type":"ContainerStarted","Data":"50c95bbed1c2660c0a8ec709a8cd8512b6b45328b609026ca6bd4a12460e87fa"} Dec 03 16:40:30 crc kubenswrapper[4998]: I1203 16:40:30.907985 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" podStartSLOduration=2.322414764 podStartE2EDuration="2.907950146s" podCreationTimestamp="2025-12-03 16:40:28 +0000 UTC" firstStartedPulling="2025-12-03 16:40:29.811807321 +0000 UTC m=+2208.423507544" lastFinishedPulling="2025-12-03 16:40:30.397342693 +0000 UTC m=+2209.009042926" observedRunningTime="2025-12-03 16:40:30.906783957 +0000 UTC m=+2209.518484180" watchObservedRunningTime="2025-12-03 16:40:30.907950146 +0000 UTC m=+2209.519650369" Dec 03 16:40:35 crc kubenswrapper[4998]: I1203 16:40:35.677953 4998 scope.go:117] "RemoveContainer" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" Dec 03 16:40:35 crc kubenswrapper[4998]: E1203 16:40:35.678795 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:40:49 crc kubenswrapper[4998]: I1203 16:40:49.679104 4998 scope.go:117] "RemoveContainer" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" Dec 03 16:40:49 crc kubenswrapper[4998]: E1203 16:40:49.680355 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:41:04 crc kubenswrapper[4998]: I1203 16:41:04.680324 4998 scope.go:117] "RemoveContainer" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" Dec 03 16:41:04 crc kubenswrapper[4998]: E1203 16:41:04.682072 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:41:13 crc kubenswrapper[4998]: I1203 16:41:13.367125 4998 generic.go:334] "Generic (PLEG): container finished" podID="c5cebbf5-5350-42c2-8908-8a508995a72c" containerID="50c95bbed1c2660c0a8ec709a8cd8512b6b45328b609026ca6bd4a12460e87fa" exitCode=0 Dec 03 16:41:13 crc kubenswrapper[4998]: I1203 16:41:13.367250 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" event={"ID":"c5cebbf5-5350-42c2-8908-8a508995a72c","Type":"ContainerDied","Data":"50c95bbed1c2660c0a8ec709a8cd8512b6b45328b609026ca6bd4a12460e87fa"} Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.885968 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.939082 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"c5cebbf5-5350-42c2-8908-8a508995a72c\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.939463 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-ssh-key\") pod \"c5cebbf5-5350-42c2-8908-8a508995a72c\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.939537 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thzwb\" (UniqueName: \"kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-kube-api-access-thzwb\") pod \"c5cebbf5-5350-42c2-8908-8a508995a72c\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.939599 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-libvirt-combined-ca-bundle\") pod \"c5cebbf5-5350-42c2-8908-8a508995a72c\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.939667 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"c5cebbf5-5350-42c2-8908-8a508995a72c\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.939711 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-ovn-combined-ca-bundle\") pod \"c5cebbf5-5350-42c2-8908-8a508995a72c\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.939880 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"c5cebbf5-5350-42c2-8908-8a508995a72c\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.939922 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-repo-setup-combined-ca-bundle\") pod \"c5cebbf5-5350-42c2-8908-8a508995a72c\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.939984 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-telemetry-combined-ca-bundle\") pod \"c5cebbf5-5350-42c2-8908-8a508995a72c\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.940029 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-bootstrap-combined-ca-bundle\") pod \"c5cebbf5-5350-42c2-8908-8a508995a72c\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.940057 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-neutron-metadata-combined-ca-bundle\") pod \"c5cebbf5-5350-42c2-8908-8a508995a72c\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.940085 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-nova-combined-ca-bundle\") pod \"c5cebbf5-5350-42c2-8908-8a508995a72c\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.940128 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"c5cebbf5-5350-42c2-8908-8a508995a72c\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.940191 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-inventory\") pod \"c5cebbf5-5350-42c2-8908-8a508995a72c\" (UID: \"c5cebbf5-5350-42c2-8908-8a508995a72c\") " Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.945413 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "c5cebbf5-5350-42c2-8908-8a508995a72c" (UID: "c5cebbf5-5350-42c2-8908-8a508995a72c"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.945481 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "c5cebbf5-5350-42c2-8908-8a508995a72c" (UID: "c5cebbf5-5350-42c2-8908-8a508995a72c"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.945517 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "c5cebbf5-5350-42c2-8908-8a508995a72c" (UID: "c5cebbf5-5350-42c2-8908-8a508995a72c"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.945672 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "c5cebbf5-5350-42c2-8908-8a508995a72c" (UID: "c5cebbf5-5350-42c2-8908-8a508995a72c"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.946222 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "c5cebbf5-5350-42c2-8908-8a508995a72c" (UID: "c5cebbf5-5350-42c2-8908-8a508995a72c"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.947575 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "c5cebbf5-5350-42c2-8908-8a508995a72c" (UID: "c5cebbf5-5350-42c2-8908-8a508995a72c"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.947593 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "c5cebbf5-5350-42c2-8908-8a508995a72c" (UID: "c5cebbf5-5350-42c2-8908-8a508995a72c"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.948854 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-kube-api-access-thzwb" (OuterVolumeSpecName: "kube-api-access-thzwb") pod "c5cebbf5-5350-42c2-8908-8a508995a72c" (UID: "c5cebbf5-5350-42c2-8908-8a508995a72c"). InnerVolumeSpecName "kube-api-access-thzwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.949387 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "c5cebbf5-5350-42c2-8908-8a508995a72c" (UID: "c5cebbf5-5350-42c2-8908-8a508995a72c"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.949850 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "c5cebbf5-5350-42c2-8908-8a508995a72c" (UID: "c5cebbf5-5350-42c2-8908-8a508995a72c"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.949909 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "c5cebbf5-5350-42c2-8908-8a508995a72c" (UID: "c5cebbf5-5350-42c2-8908-8a508995a72c"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.956703 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "c5cebbf5-5350-42c2-8908-8a508995a72c" (UID: "c5cebbf5-5350-42c2-8908-8a508995a72c"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.980565 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-inventory" (OuterVolumeSpecName: "inventory") pod "c5cebbf5-5350-42c2-8908-8a508995a72c" (UID: "c5cebbf5-5350-42c2-8908-8a508995a72c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:41:14 crc kubenswrapper[4998]: I1203 16:41:14.981518 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c5cebbf5-5350-42c2-8908-8a508995a72c" (UID: "c5cebbf5-5350-42c2-8908-8a508995a72c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.042323 4998 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.042363 4998 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.042383 4998 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.042396 4998 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.042408 4998 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.042421 4998 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.042434 4998 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.042446 4998 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.042457 4998 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.042469 4998 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.042479 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thzwb\" (UniqueName: \"kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-kube-api-access-thzwb\") on node \"crc\" DevicePath \"\"" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.042490 4998 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.042501 4998 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c5cebbf5-5350-42c2-8908-8a508995a72c-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.042513 4998 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5cebbf5-5350-42c2-8908-8a508995a72c-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.434991 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" event={"ID":"c5cebbf5-5350-42c2-8908-8a508995a72c","Type":"ContainerDied","Data":"bc206d8364647be875a9f54bd835d6204247b03a1fdb1e28236e77c03d6b4ede"} Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.435072 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc206d8364647be875a9f54bd835d6204247b03a1fdb1e28236e77c03d6b4ede" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.435118 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.693565 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk"] Dec 03 16:41:15 crc kubenswrapper[4998]: E1203 16:41:15.694280 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5cebbf5-5350-42c2-8908-8a508995a72c" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.694303 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5cebbf5-5350-42c2-8908-8a508995a72c" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.694590 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5cebbf5-5350-42c2-8908-8a508995a72c" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.695500 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.698394 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.699431 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kcj9x" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.699896 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.700097 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.700204 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.701707 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk"] Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.762202 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx2v8\" (UniqueName: \"kubernetes.io/projected/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-kube-api-access-zx2v8\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lbwsk\" (UID: \"6b9ea60c-d8a5-47a8-91dd-996ee873d6df\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.762251 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lbwsk\" (UID: \"6b9ea60c-d8a5-47a8-91dd-996ee873d6df\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.762283 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lbwsk\" (UID: \"6b9ea60c-d8a5-47a8-91dd-996ee873d6df\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.762346 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lbwsk\" (UID: \"6b9ea60c-d8a5-47a8-91dd-996ee873d6df\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.762442 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lbwsk\" (UID: \"6b9ea60c-d8a5-47a8-91dd-996ee873d6df\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.864108 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lbwsk\" (UID: \"6b9ea60c-d8a5-47a8-91dd-996ee873d6df\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.864203 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lbwsk\" (UID: \"6b9ea60c-d8a5-47a8-91dd-996ee873d6df\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.864293 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx2v8\" (UniqueName: \"kubernetes.io/projected/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-kube-api-access-zx2v8\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lbwsk\" (UID: \"6b9ea60c-d8a5-47a8-91dd-996ee873d6df\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.864310 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lbwsk\" (UID: \"6b9ea60c-d8a5-47a8-91dd-996ee873d6df\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.864330 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lbwsk\" (UID: \"6b9ea60c-d8a5-47a8-91dd-996ee873d6df\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.865205 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lbwsk\" (UID: \"6b9ea60c-d8a5-47a8-91dd-996ee873d6df\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.871156 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lbwsk\" (UID: \"6b9ea60c-d8a5-47a8-91dd-996ee873d6df\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.871318 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lbwsk\" (UID: \"6b9ea60c-d8a5-47a8-91dd-996ee873d6df\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.872307 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lbwsk\" (UID: \"6b9ea60c-d8a5-47a8-91dd-996ee873d6df\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk" Dec 03 16:41:15 crc kubenswrapper[4998]: I1203 16:41:15.902666 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx2v8\" (UniqueName: \"kubernetes.io/projected/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-kube-api-access-zx2v8\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-lbwsk\" (UID: \"6b9ea60c-d8a5-47a8-91dd-996ee873d6df\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk" Dec 03 16:41:16 crc kubenswrapper[4998]: I1203 16:41:16.015053 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk" Dec 03 16:41:16 crc kubenswrapper[4998]: I1203 16:41:16.567864 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk"] Dec 03 16:41:16 crc kubenswrapper[4998]: I1203 16:41:16.592044 4998 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 16:41:17 crc kubenswrapper[4998]: I1203 16:41:17.458287 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk" event={"ID":"6b9ea60c-d8a5-47a8-91dd-996ee873d6df","Type":"ContainerStarted","Data":"ecbf605535a47c74ca7b6a6cf33f79afbaecf11da08a0c527c70551b51dba114"} Dec 03 16:41:17 crc kubenswrapper[4998]: I1203 16:41:17.677668 4998 scope.go:117] "RemoveContainer" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" Dec 03 16:41:17 crc kubenswrapper[4998]: E1203 16:41:17.678329 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:41:18 crc kubenswrapper[4998]: I1203 16:41:18.467205 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk" event={"ID":"6b9ea60c-d8a5-47a8-91dd-996ee873d6df","Type":"ContainerStarted","Data":"aa031fd27d8c31fa11727fccbf2bdde56b967b91a928e9ef1d97f129c747e3c7"} Dec 03 16:41:18 crc kubenswrapper[4998]: I1203 16:41:18.513266 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk" podStartSLOduration=2.691845599 podStartE2EDuration="3.513239015s" podCreationTimestamp="2025-12-03 16:41:15 +0000 UTC" firstStartedPulling="2025-12-03 16:41:16.591747044 +0000 UTC m=+2255.203447287" lastFinishedPulling="2025-12-03 16:41:17.41314045 +0000 UTC m=+2256.024840703" observedRunningTime="2025-12-03 16:41:18.489722505 +0000 UTC m=+2257.101422758" watchObservedRunningTime="2025-12-03 16:41:18.513239015 +0000 UTC m=+2257.124939238" Dec 03 16:41:32 crc kubenswrapper[4998]: I1203 16:41:32.678374 4998 scope.go:117] "RemoveContainer" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" Dec 03 16:41:32 crc kubenswrapper[4998]: E1203 16:41:32.679371 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:41:44 crc kubenswrapper[4998]: I1203 16:41:44.679716 4998 scope.go:117] "RemoveContainer" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" Dec 03 16:41:44 crc kubenswrapper[4998]: E1203 16:41:44.681067 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:41:56 crc kubenswrapper[4998]: I1203 16:41:56.679219 4998 scope.go:117] "RemoveContainer" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" Dec 03 16:41:56 crc kubenswrapper[4998]: E1203 16:41:56.680474 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:42:07 crc kubenswrapper[4998]: I1203 16:42:07.678294 4998 scope.go:117] "RemoveContainer" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" Dec 03 16:42:07 crc kubenswrapper[4998]: E1203 16:42:07.679049 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:42:21 crc kubenswrapper[4998]: I1203 16:42:21.691595 4998 scope.go:117] "RemoveContainer" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" Dec 03 16:42:21 crc kubenswrapper[4998]: E1203 16:42:21.692771 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:42:30 crc kubenswrapper[4998]: I1203 16:42:30.263399 4998 generic.go:334] "Generic (PLEG): container finished" podID="6b9ea60c-d8a5-47a8-91dd-996ee873d6df" containerID="aa031fd27d8c31fa11727fccbf2bdde56b967b91a928e9ef1d97f129c747e3c7" exitCode=0 Dec 03 16:42:30 crc kubenswrapper[4998]: I1203 16:42:30.263522 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk" event={"ID":"6b9ea60c-d8a5-47a8-91dd-996ee873d6df","Type":"ContainerDied","Data":"aa031fd27d8c31fa11727fccbf2bdde56b967b91a928e9ef1d97f129c747e3c7"} Dec 03 16:42:31 crc kubenswrapper[4998]: I1203 16:42:31.735229 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk" Dec 03 16:42:31 crc kubenswrapper[4998]: I1203 16:42:31.837262 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-ssh-key\") pod \"6b9ea60c-d8a5-47a8-91dd-996ee873d6df\" (UID: \"6b9ea60c-d8a5-47a8-91dd-996ee873d6df\") " Dec 03 16:42:31 crc kubenswrapper[4998]: I1203 16:42:31.837383 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-ovn-combined-ca-bundle\") pod \"6b9ea60c-d8a5-47a8-91dd-996ee873d6df\" (UID: \"6b9ea60c-d8a5-47a8-91dd-996ee873d6df\") " Dec 03 16:42:31 crc kubenswrapper[4998]: I1203 16:42:31.837427 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-ovncontroller-config-0\") pod \"6b9ea60c-d8a5-47a8-91dd-996ee873d6df\" (UID: \"6b9ea60c-d8a5-47a8-91dd-996ee873d6df\") " Dec 03 16:42:31 crc kubenswrapper[4998]: I1203 16:42:31.837563 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-inventory\") pod \"6b9ea60c-d8a5-47a8-91dd-996ee873d6df\" (UID: \"6b9ea60c-d8a5-47a8-91dd-996ee873d6df\") " Dec 03 16:42:31 crc kubenswrapper[4998]: I1203 16:42:31.837887 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zx2v8\" (UniqueName: \"kubernetes.io/projected/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-kube-api-access-zx2v8\") pod \"6b9ea60c-d8a5-47a8-91dd-996ee873d6df\" (UID: \"6b9ea60c-d8a5-47a8-91dd-996ee873d6df\") " Dec 03 16:42:31 crc kubenswrapper[4998]: I1203 16:42:31.843497 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "6b9ea60c-d8a5-47a8-91dd-996ee873d6df" (UID: "6b9ea60c-d8a5-47a8-91dd-996ee873d6df"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:42:31 crc kubenswrapper[4998]: I1203 16:42:31.843964 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-kube-api-access-zx2v8" (OuterVolumeSpecName: "kube-api-access-zx2v8") pod "6b9ea60c-d8a5-47a8-91dd-996ee873d6df" (UID: "6b9ea60c-d8a5-47a8-91dd-996ee873d6df"). InnerVolumeSpecName "kube-api-access-zx2v8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:42:31 crc kubenswrapper[4998]: I1203 16:42:31.865368 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "6b9ea60c-d8a5-47a8-91dd-996ee873d6df" (UID: "6b9ea60c-d8a5-47a8-91dd-996ee873d6df"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:42:31 crc kubenswrapper[4998]: I1203 16:42:31.866245 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-inventory" (OuterVolumeSpecName: "inventory") pod "6b9ea60c-d8a5-47a8-91dd-996ee873d6df" (UID: "6b9ea60c-d8a5-47a8-91dd-996ee873d6df"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:42:31 crc kubenswrapper[4998]: I1203 16:42:31.873081 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6b9ea60c-d8a5-47a8-91dd-996ee873d6df" (UID: "6b9ea60c-d8a5-47a8-91dd-996ee873d6df"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:42:31 crc kubenswrapper[4998]: I1203 16:42:31.940489 4998 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 16:42:31 crc kubenswrapper[4998]: I1203 16:42:31.940536 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zx2v8\" (UniqueName: \"kubernetes.io/projected/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-kube-api-access-zx2v8\") on node \"crc\" DevicePath \"\"" Dec 03 16:42:31 crc kubenswrapper[4998]: I1203 16:42:31.940552 4998 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 16:42:31 crc kubenswrapper[4998]: I1203 16:42:31.940567 4998 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:42:31 crc kubenswrapper[4998]: I1203 16:42:31.940579 4998 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6b9ea60c-d8a5-47a8-91dd-996ee873d6df-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.288827 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk" event={"ID":"6b9ea60c-d8a5-47a8-91dd-996ee873d6df","Type":"ContainerDied","Data":"ecbf605535a47c74ca7b6a6cf33f79afbaecf11da08a0c527c70551b51dba114"} Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.288883 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ecbf605535a47c74ca7b6a6cf33f79afbaecf11da08a0c527c70551b51dba114" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.288935 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-lbwsk" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.383238 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8"] Dec 03 16:42:32 crc kubenswrapper[4998]: E1203 16:42:32.383652 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b9ea60c-d8a5-47a8-91dd-996ee873d6df" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.383668 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b9ea60c-d8a5-47a8-91dd-996ee873d6df" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.383930 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b9ea60c-d8a5-47a8-91dd-996ee873d6df" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.384572 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.389368 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.389507 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.389637 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.393688 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.393986 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.394130 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kcj9x" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.401862 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8"] Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.448403 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8\" (UID: \"d93c0246-c19d-45b5-bee1-549fdf19cacd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.448466 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8\" (UID: \"d93c0246-c19d-45b5-bee1-549fdf19cacd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.448534 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8\" (UID: \"d93c0246-c19d-45b5-bee1-549fdf19cacd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.448574 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8\" (UID: \"d93c0246-c19d-45b5-bee1-549fdf19cacd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.448617 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8\" (UID: \"d93c0246-c19d-45b5-bee1-549fdf19cacd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.448732 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cs24t\" (UniqueName: \"kubernetes.io/projected/d93c0246-c19d-45b5-bee1-549fdf19cacd-kube-api-access-cs24t\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8\" (UID: \"d93c0246-c19d-45b5-bee1-549fdf19cacd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.550079 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8\" (UID: \"d93c0246-c19d-45b5-bee1-549fdf19cacd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.550145 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8\" (UID: \"d93c0246-c19d-45b5-bee1-549fdf19cacd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.550208 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8\" (UID: \"d93c0246-c19d-45b5-bee1-549fdf19cacd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.550245 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8\" (UID: \"d93c0246-c19d-45b5-bee1-549fdf19cacd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.550287 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8\" (UID: \"d93c0246-c19d-45b5-bee1-549fdf19cacd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.550371 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cs24t\" (UniqueName: \"kubernetes.io/projected/d93c0246-c19d-45b5-bee1-549fdf19cacd-kube-api-access-cs24t\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8\" (UID: \"d93c0246-c19d-45b5-bee1-549fdf19cacd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.553823 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8\" (UID: \"d93c0246-c19d-45b5-bee1-549fdf19cacd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.554040 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8\" (UID: \"d93c0246-c19d-45b5-bee1-549fdf19cacd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.554280 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8\" (UID: \"d93c0246-c19d-45b5-bee1-549fdf19cacd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.554333 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8\" (UID: \"d93c0246-c19d-45b5-bee1-549fdf19cacd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.554765 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8\" (UID: \"d93c0246-c19d-45b5-bee1-549fdf19cacd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.573296 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cs24t\" (UniqueName: \"kubernetes.io/projected/d93c0246-c19d-45b5-bee1-549fdf19cacd-kube-api-access-cs24t\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8\" (UID: \"d93c0246-c19d-45b5-bee1-549fdf19cacd\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" Dec 03 16:42:32 crc kubenswrapper[4998]: I1203 16:42:32.701237 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" Dec 03 16:42:33 crc kubenswrapper[4998]: I1203 16:42:33.248534 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8"] Dec 03 16:42:33 crc kubenswrapper[4998]: I1203 16:42:33.297251 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" event={"ID":"d93c0246-c19d-45b5-bee1-549fdf19cacd","Type":"ContainerStarted","Data":"327e2688f90fcfe4505e996f28e9ee468734c4b0dce2ec99a1fc60a8cd0de434"} Dec 03 16:42:34 crc kubenswrapper[4998]: I1203 16:42:34.310533 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" event={"ID":"d93c0246-c19d-45b5-bee1-549fdf19cacd","Type":"ContainerStarted","Data":"887ff60b719d4909e2b4f0a02c4310fd3387da49a505f6a2cca1d6775a5469d0"} Dec 03 16:42:34 crc kubenswrapper[4998]: I1203 16:42:34.337528 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" podStartSLOduration=1.872431582 podStartE2EDuration="2.337511084s" podCreationTimestamp="2025-12-03 16:42:32 +0000 UTC" firstStartedPulling="2025-12-03 16:42:33.257299618 +0000 UTC m=+2331.868999841" lastFinishedPulling="2025-12-03 16:42:33.72237912 +0000 UTC m=+2332.334079343" observedRunningTime="2025-12-03 16:42:34.329132634 +0000 UTC m=+2332.940832857" watchObservedRunningTime="2025-12-03 16:42:34.337511084 +0000 UTC m=+2332.949211307" Dec 03 16:42:36 crc kubenswrapper[4998]: I1203 16:42:36.677513 4998 scope.go:117] "RemoveContainer" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" Dec 03 16:42:36 crc kubenswrapper[4998]: E1203 16:42:36.678516 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:42:50 crc kubenswrapper[4998]: I1203 16:42:50.678659 4998 scope.go:117] "RemoveContainer" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" Dec 03 16:42:50 crc kubenswrapper[4998]: E1203 16:42:50.679613 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:43:03 crc kubenswrapper[4998]: I1203 16:43:03.678967 4998 scope.go:117] "RemoveContainer" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" Dec 03 16:43:03 crc kubenswrapper[4998]: E1203 16:43:03.680245 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:43:17 crc kubenswrapper[4998]: I1203 16:43:17.678990 4998 scope.go:117] "RemoveContainer" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" Dec 03 16:43:17 crc kubenswrapper[4998]: E1203 16:43:17.680123 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:43:27 crc kubenswrapper[4998]: I1203 16:43:27.834793 4998 generic.go:334] "Generic (PLEG): container finished" podID="d93c0246-c19d-45b5-bee1-549fdf19cacd" containerID="887ff60b719d4909e2b4f0a02c4310fd3387da49a505f6a2cca1d6775a5469d0" exitCode=0 Dec 03 16:43:27 crc kubenswrapper[4998]: I1203 16:43:27.834793 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" event={"ID":"d93c0246-c19d-45b5-bee1-549fdf19cacd","Type":"ContainerDied","Data":"887ff60b719d4909e2b4f0a02c4310fd3387da49a505f6a2cca1d6775a5469d0"} Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.271391 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.312205 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-inventory\") pod \"d93c0246-c19d-45b5-bee1-549fdf19cacd\" (UID: \"d93c0246-c19d-45b5-bee1-549fdf19cacd\") " Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.312418 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-neutron-metadata-combined-ca-bundle\") pod \"d93c0246-c19d-45b5-bee1-549fdf19cacd\" (UID: \"d93c0246-c19d-45b5-bee1-549fdf19cacd\") " Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.312464 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cs24t\" (UniqueName: \"kubernetes.io/projected/d93c0246-c19d-45b5-bee1-549fdf19cacd-kube-api-access-cs24t\") pod \"d93c0246-c19d-45b5-bee1-549fdf19cacd\" (UID: \"d93c0246-c19d-45b5-bee1-549fdf19cacd\") " Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.312523 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-ssh-key\") pod \"d93c0246-c19d-45b5-bee1-549fdf19cacd\" (UID: \"d93c0246-c19d-45b5-bee1-549fdf19cacd\") " Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.312607 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-nova-metadata-neutron-config-0\") pod \"d93c0246-c19d-45b5-bee1-549fdf19cacd\" (UID: \"d93c0246-c19d-45b5-bee1-549fdf19cacd\") " Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.312641 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-neutron-ovn-metadata-agent-neutron-config-0\") pod \"d93c0246-c19d-45b5-bee1-549fdf19cacd\" (UID: \"d93c0246-c19d-45b5-bee1-549fdf19cacd\") " Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.320613 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "d93c0246-c19d-45b5-bee1-549fdf19cacd" (UID: "d93c0246-c19d-45b5-bee1-549fdf19cacd"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.320818 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d93c0246-c19d-45b5-bee1-549fdf19cacd-kube-api-access-cs24t" (OuterVolumeSpecName: "kube-api-access-cs24t") pod "d93c0246-c19d-45b5-bee1-549fdf19cacd" (UID: "d93c0246-c19d-45b5-bee1-549fdf19cacd"). InnerVolumeSpecName "kube-api-access-cs24t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.346647 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d93c0246-c19d-45b5-bee1-549fdf19cacd" (UID: "d93c0246-c19d-45b5-bee1-549fdf19cacd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.347258 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "d93c0246-c19d-45b5-bee1-549fdf19cacd" (UID: "d93c0246-c19d-45b5-bee1-549fdf19cacd"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.350965 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-inventory" (OuterVolumeSpecName: "inventory") pod "d93c0246-c19d-45b5-bee1-549fdf19cacd" (UID: "d93c0246-c19d-45b5-bee1-549fdf19cacd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.355966 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "d93c0246-c19d-45b5-bee1-549fdf19cacd" (UID: "d93c0246-c19d-45b5-bee1-549fdf19cacd"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.414935 4998 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.414983 4998 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.415006 4998 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.415026 4998 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.415047 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cs24t\" (UniqueName: \"kubernetes.io/projected/d93c0246-c19d-45b5-bee1-549fdf19cacd-kube-api-access-cs24t\") on node \"crc\" DevicePath \"\"" Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.415067 4998 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d93c0246-c19d-45b5-bee1-549fdf19cacd-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.857788 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" event={"ID":"d93c0246-c19d-45b5-bee1-549fdf19cacd","Type":"ContainerDied","Data":"327e2688f90fcfe4505e996f28e9ee468734c4b0dce2ec99a1fc60a8cd0de434"} Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.858130 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="327e2688f90fcfe4505e996f28e9ee468734c4b0dce2ec99a1fc60a8cd0de434" Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.858076 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8" Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.942049 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q"] Dec 03 16:43:29 crc kubenswrapper[4998]: E1203 16:43:29.942433 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d93c0246-c19d-45b5-bee1-549fdf19cacd" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.942450 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="d93c0246-c19d-45b5-bee1-549fdf19cacd" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.942660 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="d93c0246-c19d-45b5-bee1-549fdf19cacd" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.943631 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q" Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.945893 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.947945 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.947994 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kcj9x" Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.948231 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.948396 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 16:43:29 crc kubenswrapper[4998]: I1203 16:43:29.957086 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q"] Dec 03 16:43:30 crc kubenswrapper[4998]: I1203 16:43:30.126817 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q\" (UID: \"d23ecf9c-0347-4a37-ae82-c7b24d5b9575\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q" Dec 03 16:43:30 crc kubenswrapper[4998]: I1203 16:43:30.126975 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fp2n\" (UniqueName: \"kubernetes.io/projected/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-kube-api-access-9fp2n\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q\" (UID: \"d23ecf9c-0347-4a37-ae82-c7b24d5b9575\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q" Dec 03 16:43:30 crc kubenswrapper[4998]: I1203 16:43:30.127067 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q\" (UID: \"d23ecf9c-0347-4a37-ae82-c7b24d5b9575\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q" Dec 03 16:43:30 crc kubenswrapper[4998]: I1203 16:43:30.127175 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q\" (UID: \"d23ecf9c-0347-4a37-ae82-c7b24d5b9575\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q" Dec 03 16:43:30 crc kubenswrapper[4998]: I1203 16:43:30.127213 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q\" (UID: \"d23ecf9c-0347-4a37-ae82-c7b24d5b9575\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q" Dec 03 16:43:30 crc kubenswrapper[4998]: I1203 16:43:30.229459 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q\" (UID: \"d23ecf9c-0347-4a37-ae82-c7b24d5b9575\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q" Dec 03 16:43:30 crc kubenswrapper[4998]: I1203 16:43:30.229580 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fp2n\" (UniqueName: \"kubernetes.io/projected/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-kube-api-access-9fp2n\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q\" (UID: \"d23ecf9c-0347-4a37-ae82-c7b24d5b9575\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q" Dec 03 16:43:30 crc kubenswrapper[4998]: I1203 16:43:30.229631 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q\" (UID: \"d23ecf9c-0347-4a37-ae82-c7b24d5b9575\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q" Dec 03 16:43:30 crc kubenswrapper[4998]: I1203 16:43:30.229675 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q\" (UID: \"d23ecf9c-0347-4a37-ae82-c7b24d5b9575\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q" Dec 03 16:43:30 crc kubenswrapper[4998]: I1203 16:43:30.229713 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q\" (UID: \"d23ecf9c-0347-4a37-ae82-c7b24d5b9575\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q" Dec 03 16:43:30 crc kubenswrapper[4998]: I1203 16:43:30.234360 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q\" (UID: \"d23ecf9c-0347-4a37-ae82-c7b24d5b9575\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q" Dec 03 16:43:30 crc kubenswrapper[4998]: I1203 16:43:30.242453 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q\" (UID: \"d23ecf9c-0347-4a37-ae82-c7b24d5b9575\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q" Dec 03 16:43:30 crc kubenswrapper[4998]: I1203 16:43:30.242516 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q\" (UID: \"d23ecf9c-0347-4a37-ae82-c7b24d5b9575\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q" Dec 03 16:43:30 crc kubenswrapper[4998]: I1203 16:43:30.242918 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q\" (UID: \"d23ecf9c-0347-4a37-ae82-c7b24d5b9575\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q" Dec 03 16:43:30 crc kubenswrapper[4998]: I1203 16:43:30.255393 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fp2n\" (UniqueName: \"kubernetes.io/projected/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-kube-api-access-9fp2n\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q\" (UID: \"d23ecf9c-0347-4a37-ae82-c7b24d5b9575\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q" Dec 03 16:43:30 crc kubenswrapper[4998]: I1203 16:43:30.262995 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q" Dec 03 16:43:30 crc kubenswrapper[4998]: I1203 16:43:30.841229 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q"] Dec 03 16:43:30 crc kubenswrapper[4998]: I1203 16:43:30.869369 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q" event={"ID":"d23ecf9c-0347-4a37-ae82-c7b24d5b9575","Type":"ContainerStarted","Data":"53be1987a9081f6cf510961134fb3aa23cb44e7858a06343d7a0bfcbe1a127ed"} Dec 03 16:43:31 crc kubenswrapper[4998]: I1203 16:43:31.892179 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q" event={"ID":"d23ecf9c-0347-4a37-ae82-c7b24d5b9575","Type":"ContainerStarted","Data":"d632ab7fddf239b15629f7e135d77abd3b1758ee543f3caead3ec023967fd3be"} Dec 03 16:43:32 crc kubenswrapper[4998]: I1203 16:43:32.677481 4998 scope.go:117] "RemoveContainer" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" Dec 03 16:43:32 crc kubenswrapper[4998]: E1203 16:43:32.677894 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:43:46 crc kubenswrapper[4998]: I1203 16:43:46.678730 4998 scope.go:117] "RemoveContainer" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" Dec 03 16:43:46 crc kubenswrapper[4998]: E1203 16:43:46.679659 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:43:59 crc kubenswrapper[4998]: I1203 16:43:59.678355 4998 scope.go:117] "RemoveContainer" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" Dec 03 16:43:59 crc kubenswrapper[4998]: E1203 16:43:59.679717 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:44:14 crc kubenswrapper[4998]: I1203 16:44:14.677936 4998 scope.go:117] "RemoveContainer" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" Dec 03 16:44:14 crc kubenswrapper[4998]: E1203 16:44:14.678703 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:44:26 crc kubenswrapper[4998]: I1203 16:44:26.677499 4998 scope.go:117] "RemoveContainer" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" Dec 03 16:44:26 crc kubenswrapper[4998]: E1203 16:44:26.679663 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:44:37 crc kubenswrapper[4998]: I1203 16:44:37.677848 4998 scope.go:117] "RemoveContainer" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" Dec 03 16:44:37 crc kubenswrapper[4998]: E1203 16:44:37.678833 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:44:51 crc kubenswrapper[4998]: I1203 16:44:51.690075 4998 scope.go:117] "RemoveContainer" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" Dec 03 16:44:51 crc kubenswrapper[4998]: E1203 16:44:51.691126 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:45:00 crc kubenswrapper[4998]: I1203 16:45:00.139406 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q" podStartSLOduration=90.604718171 podStartE2EDuration="1m31.139385138s" podCreationTimestamp="2025-12-03 16:43:29 +0000 UTC" firstStartedPulling="2025-12-03 16:43:30.843498132 +0000 UTC m=+2389.455198355" lastFinishedPulling="2025-12-03 16:43:31.378165099 +0000 UTC m=+2389.989865322" observedRunningTime="2025-12-03 16:43:31.912333673 +0000 UTC m=+2390.524033936" watchObservedRunningTime="2025-12-03 16:45:00.139385138 +0000 UTC m=+2478.751085361" Dec 03 16:45:00 crc kubenswrapper[4998]: I1203 16:45:00.152155 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413005-4z5cj"] Dec 03 16:45:00 crc kubenswrapper[4998]: I1203 16:45:00.154402 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-4z5cj" Dec 03 16:45:00 crc kubenswrapper[4998]: I1203 16:45:00.157336 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 16:45:00 crc kubenswrapper[4998]: I1203 16:45:00.158018 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 16:45:00 crc kubenswrapper[4998]: I1203 16:45:00.162031 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413005-4z5cj"] Dec 03 16:45:00 crc kubenswrapper[4998]: I1203 16:45:00.258175 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51fd9c20-61fa-4b51-913c-6c6745d881e0-config-volume\") pod \"collect-profiles-29413005-4z5cj\" (UID: \"51fd9c20-61fa-4b51-913c-6c6745d881e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-4z5cj" Dec 03 16:45:00 crc kubenswrapper[4998]: I1203 16:45:00.258813 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/51fd9c20-61fa-4b51-913c-6c6745d881e0-secret-volume\") pod \"collect-profiles-29413005-4z5cj\" (UID: \"51fd9c20-61fa-4b51-913c-6c6745d881e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-4z5cj" Dec 03 16:45:00 crc kubenswrapper[4998]: I1203 16:45:00.258957 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpfmg\" (UniqueName: \"kubernetes.io/projected/51fd9c20-61fa-4b51-913c-6c6745d881e0-kube-api-access-lpfmg\") pod \"collect-profiles-29413005-4z5cj\" (UID: \"51fd9c20-61fa-4b51-913c-6c6745d881e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-4z5cj" Dec 03 16:45:00 crc kubenswrapper[4998]: I1203 16:45:00.360914 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/51fd9c20-61fa-4b51-913c-6c6745d881e0-secret-volume\") pod \"collect-profiles-29413005-4z5cj\" (UID: \"51fd9c20-61fa-4b51-913c-6c6745d881e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-4z5cj" Dec 03 16:45:00 crc kubenswrapper[4998]: I1203 16:45:00.360997 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpfmg\" (UniqueName: \"kubernetes.io/projected/51fd9c20-61fa-4b51-913c-6c6745d881e0-kube-api-access-lpfmg\") pod \"collect-profiles-29413005-4z5cj\" (UID: \"51fd9c20-61fa-4b51-913c-6c6745d881e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-4z5cj" Dec 03 16:45:00 crc kubenswrapper[4998]: I1203 16:45:00.361087 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51fd9c20-61fa-4b51-913c-6c6745d881e0-config-volume\") pod \"collect-profiles-29413005-4z5cj\" (UID: \"51fd9c20-61fa-4b51-913c-6c6745d881e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-4z5cj" Dec 03 16:45:00 crc kubenswrapper[4998]: I1203 16:45:00.361995 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51fd9c20-61fa-4b51-913c-6c6745d881e0-config-volume\") pod \"collect-profiles-29413005-4z5cj\" (UID: \"51fd9c20-61fa-4b51-913c-6c6745d881e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-4z5cj" Dec 03 16:45:00 crc kubenswrapper[4998]: I1203 16:45:00.370489 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/51fd9c20-61fa-4b51-913c-6c6745d881e0-secret-volume\") pod \"collect-profiles-29413005-4z5cj\" (UID: \"51fd9c20-61fa-4b51-913c-6c6745d881e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-4z5cj" Dec 03 16:45:00 crc kubenswrapper[4998]: I1203 16:45:00.382387 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpfmg\" (UniqueName: \"kubernetes.io/projected/51fd9c20-61fa-4b51-913c-6c6745d881e0-kube-api-access-lpfmg\") pod \"collect-profiles-29413005-4z5cj\" (UID: \"51fd9c20-61fa-4b51-913c-6c6745d881e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-4z5cj" Dec 03 16:45:00 crc kubenswrapper[4998]: I1203 16:45:00.484075 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-4z5cj" Dec 03 16:45:00 crc kubenswrapper[4998]: I1203 16:45:00.939893 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413005-4z5cj"] Dec 03 16:45:01 crc kubenswrapper[4998]: I1203 16:45:01.799196 4998 generic.go:334] "Generic (PLEG): container finished" podID="51fd9c20-61fa-4b51-913c-6c6745d881e0" containerID="6bfc46eb16ba5179dfa693bfce1486f9ef722916c8726a03073ca063f8d44e18" exitCode=0 Dec 03 16:45:01 crc kubenswrapper[4998]: I1203 16:45:01.799447 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-4z5cj" event={"ID":"51fd9c20-61fa-4b51-913c-6c6745d881e0","Type":"ContainerDied","Data":"6bfc46eb16ba5179dfa693bfce1486f9ef722916c8726a03073ca063f8d44e18"} Dec 03 16:45:01 crc kubenswrapper[4998]: I1203 16:45:01.799475 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-4z5cj" event={"ID":"51fd9c20-61fa-4b51-913c-6c6745d881e0","Type":"ContainerStarted","Data":"c7441f0b2091edbd2a151bd0b90b3e68ca2059ebf6005e6872df49ed280ff86b"} Dec 03 16:45:03 crc kubenswrapper[4998]: I1203 16:45:03.153818 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-4z5cj" Dec 03 16:45:03 crc kubenswrapper[4998]: I1203 16:45:03.215478 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpfmg\" (UniqueName: \"kubernetes.io/projected/51fd9c20-61fa-4b51-913c-6c6745d881e0-kube-api-access-lpfmg\") pod \"51fd9c20-61fa-4b51-913c-6c6745d881e0\" (UID: \"51fd9c20-61fa-4b51-913c-6c6745d881e0\") " Dec 03 16:45:03 crc kubenswrapper[4998]: I1203 16:45:03.215566 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51fd9c20-61fa-4b51-913c-6c6745d881e0-config-volume\") pod \"51fd9c20-61fa-4b51-913c-6c6745d881e0\" (UID: \"51fd9c20-61fa-4b51-913c-6c6745d881e0\") " Dec 03 16:45:03 crc kubenswrapper[4998]: I1203 16:45:03.215739 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/51fd9c20-61fa-4b51-913c-6c6745d881e0-secret-volume\") pod \"51fd9c20-61fa-4b51-913c-6c6745d881e0\" (UID: \"51fd9c20-61fa-4b51-913c-6c6745d881e0\") " Dec 03 16:45:03 crc kubenswrapper[4998]: I1203 16:45:03.216260 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51fd9c20-61fa-4b51-913c-6c6745d881e0-config-volume" (OuterVolumeSpecName: "config-volume") pod "51fd9c20-61fa-4b51-913c-6c6745d881e0" (UID: "51fd9c20-61fa-4b51-913c-6c6745d881e0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:45:03 crc kubenswrapper[4998]: I1203 16:45:03.224850 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51fd9c20-61fa-4b51-913c-6c6745d881e0-kube-api-access-lpfmg" (OuterVolumeSpecName: "kube-api-access-lpfmg") pod "51fd9c20-61fa-4b51-913c-6c6745d881e0" (UID: "51fd9c20-61fa-4b51-913c-6c6745d881e0"). InnerVolumeSpecName "kube-api-access-lpfmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:45:03 crc kubenswrapper[4998]: I1203 16:45:03.225102 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51fd9c20-61fa-4b51-913c-6c6745d881e0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "51fd9c20-61fa-4b51-913c-6c6745d881e0" (UID: "51fd9c20-61fa-4b51-913c-6c6745d881e0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:45:03 crc kubenswrapper[4998]: I1203 16:45:03.318106 4998 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/51fd9c20-61fa-4b51-913c-6c6745d881e0-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 16:45:03 crc kubenswrapper[4998]: I1203 16:45:03.318150 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpfmg\" (UniqueName: \"kubernetes.io/projected/51fd9c20-61fa-4b51-913c-6c6745d881e0-kube-api-access-lpfmg\") on node \"crc\" DevicePath \"\"" Dec 03 16:45:03 crc kubenswrapper[4998]: I1203 16:45:03.318165 4998 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51fd9c20-61fa-4b51-913c-6c6745d881e0-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 16:45:03 crc kubenswrapper[4998]: I1203 16:45:03.677875 4998 scope.go:117] "RemoveContainer" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" Dec 03 16:45:03 crc kubenswrapper[4998]: I1203 16:45:03.822094 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-4z5cj" event={"ID":"51fd9c20-61fa-4b51-913c-6c6745d881e0","Type":"ContainerDied","Data":"c7441f0b2091edbd2a151bd0b90b3e68ca2059ebf6005e6872df49ed280ff86b"} Dec 03 16:45:03 crc kubenswrapper[4998]: I1203 16:45:03.822138 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7441f0b2091edbd2a151bd0b90b3e68ca2059ebf6005e6872df49ed280ff86b" Dec 03 16:45:03 crc kubenswrapper[4998]: I1203 16:45:03.822178 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413005-4z5cj" Dec 03 16:45:04 crc kubenswrapper[4998]: I1203 16:45:04.236358 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412960-6vr8n"] Dec 03 16:45:04 crc kubenswrapper[4998]: I1203 16:45:04.246600 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412960-6vr8n"] Dec 03 16:45:04 crc kubenswrapper[4998]: I1203 16:45:04.832875 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"65e608d68983552183ea47364a0c23e120fed4872f3ce284dfdd07db6b645cc3"} Dec 03 16:45:05 crc kubenswrapper[4998]: I1203 16:45:05.716938 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6e6bbd8-c854-4708-b9b1-40884d1be6c7" path="/var/lib/kubelet/pods/a6e6bbd8-c854-4708-b9b1-40884d1be6c7/volumes" Dec 03 16:46:00 crc kubenswrapper[4998]: I1203 16:46:00.665910 4998 scope.go:117] "RemoveContainer" containerID="481b1802323e8456ca60f05e1a85214ce7a71c0cba6ebdbe39be6ec00757d272" Dec 03 16:46:11 crc kubenswrapper[4998]: I1203 16:46:11.245553 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kg9v4"] Dec 03 16:46:11 crc kubenswrapper[4998]: E1203 16:46:11.247475 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51fd9c20-61fa-4b51-913c-6c6745d881e0" containerName="collect-profiles" Dec 03 16:46:11 crc kubenswrapper[4998]: I1203 16:46:11.247558 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="51fd9c20-61fa-4b51-913c-6c6745d881e0" containerName="collect-profiles" Dec 03 16:46:11 crc kubenswrapper[4998]: I1203 16:46:11.247841 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="51fd9c20-61fa-4b51-913c-6c6745d881e0" containerName="collect-profiles" Dec 03 16:46:11 crc kubenswrapper[4998]: I1203 16:46:11.249376 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kg9v4" Dec 03 16:46:11 crc kubenswrapper[4998]: I1203 16:46:11.262697 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kg9v4"] Dec 03 16:46:11 crc kubenswrapper[4998]: I1203 16:46:11.349902 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/beac9431-5b94-4cec-b29c-7c01f9de8e59-utilities\") pod \"redhat-marketplace-kg9v4\" (UID: \"beac9431-5b94-4cec-b29c-7c01f9de8e59\") " pod="openshift-marketplace/redhat-marketplace-kg9v4" Dec 03 16:46:11 crc kubenswrapper[4998]: I1203 16:46:11.350183 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb4xz\" (UniqueName: \"kubernetes.io/projected/beac9431-5b94-4cec-b29c-7c01f9de8e59-kube-api-access-gb4xz\") pod \"redhat-marketplace-kg9v4\" (UID: \"beac9431-5b94-4cec-b29c-7c01f9de8e59\") " pod="openshift-marketplace/redhat-marketplace-kg9v4" Dec 03 16:46:11 crc kubenswrapper[4998]: I1203 16:46:11.350441 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/beac9431-5b94-4cec-b29c-7c01f9de8e59-catalog-content\") pod \"redhat-marketplace-kg9v4\" (UID: \"beac9431-5b94-4cec-b29c-7c01f9de8e59\") " pod="openshift-marketplace/redhat-marketplace-kg9v4" Dec 03 16:46:11 crc kubenswrapper[4998]: I1203 16:46:11.452491 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/beac9431-5b94-4cec-b29c-7c01f9de8e59-catalog-content\") pod \"redhat-marketplace-kg9v4\" (UID: \"beac9431-5b94-4cec-b29c-7c01f9de8e59\") " pod="openshift-marketplace/redhat-marketplace-kg9v4" Dec 03 16:46:11 crc kubenswrapper[4998]: I1203 16:46:11.452889 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/beac9431-5b94-4cec-b29c-7c01f9de8e59-utilities\") pod \"redhat-marketplace-kg9v4\" (UID: \"beac9431-5b94-4cec-b29c-7c01f9de8e59\") " pod="openshift-marketplace/redhat-marketplace-kg9v4" Dec 03 16:46:11 crc kubenswrapper[4998]: I1203 16:46:11.453001 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb4xz\" (UniqueName: \"kubernetes.io/projected/beac9431-5b94-4cec-b29c-7c01f9de8e59-kube-api-access-gb4xz\") pod \"redhat-marketplace-kg9v4\" (UID: \"beac9431-5b94-4cec-b29c-7c01f9de8e59\") " pod="openshift-marketplace/redhat-marketplace-kg9v4" Dec 03 16:46:11 crc kubenswrapper[4998]: I1203 16:46:11.453094 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/beac9431-5b94-4cec-b29c-7c01f9de8e59-catalog-content\") pod \"redhat-marketplace-kg9v4\" (UID: \"beac9431-5b94-4cec-b29c-7c01f9de8e59\") " pod="openshift-marketplace/redhat-marketplace-kg9v4" Dec 03 16:46:11 crc kubenswrapper[4998]: I1203 16:46:11.453285 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/beac9431-5b94-4cec-b29c-7c01f9de8e59-utilities\") pod \"redhat-marketplace-kg9v4\" (UID: \"beac9431-5b94-4cec-b29c-7c01f9de8e59\") " pod="openshift-marketplace/redhat-marketplace-kg9v4" Dec 03 16:46:11 crc kubenswrapper[4998]: I1203 16:46:11.473041 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb4xz\" (UniqueName: \"kubernetes.io/projected/beac9431-5b94-4cec-b29c-7c01f9de8e59-kube-api-access-gb4xz\") pod \"redhat-marketplace-kg9v4\" (UID: \"beac9431-5b94-4cec-b29c-7c01f9de8e59\") " pod="openshift-marketplace/redhat-marketplace-kg9v4" Dec 03 16:46:11 crc kubenswrapper[4998]: I1203 16:46:11.617527 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kg9v4" Dec 03 16:46:12 crc kubenswrapper[4998]: I1203 16:46:12.082892 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kg9v4"] Dec 03 16:46:12 crc kubenswrapper[4998]: I1203 16:46:12.553389 4998 generic.go:334] "Generic (PLEG): container finished" podID="beac9431-5b94-4cec-b29c-7c01f9de8e59" containerID="69d73dbecd2728b9454a688e1b923605e3d866412d7365a25e9ae4d45251c8d4" exitCode=0 Dec 03 16:46:12 crc kubenswrapper[4998]: I1203 16:46:12.553436 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kg9v4" event={"ID":"beac9431-5b94-4cec-b29c-7c01f9de8e59","Type":"ContainerDied","Data":"69d73dbecd2728b9454a688e1b923605e3d866412d7365a25e9ae4d45251c8d4"} Dec 03 16:46:12 crc kubenswrapper[4998]: I1203 16:46:12.554875 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kg9v4" event={"ID":"beac9431-5b94-4cec-b29c-7c01f9de8e59","Type":"ContainerStarted","Data":"bb8ddb712cf32426a8583ce57da5650a6e099745b1e72f70241bd71f1e4a6823"} Dec 03 16:46:13 crc kubenswrapper[4998]: I1203 16:46:13.566053 4998 generic.go:334] "Generic (PLEG): container finished" podID="beac9431-5b94-4cec-b29c-7c01f9de8e59" containerID="c58cce204fa47a5fa3a79c2c859c5daa6e562a73328ce422b322357e666cf084" exitCode=0 Dec 03 16:46:13 crc kubenswrapper[4998]: I1203 16:46:13.566099 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kg9v4" event={"ID":"beac9431-5b94-4cec-b29c-7c01f9de8e59","Type":"ContainerDied","Data":"c58cce204fa47a5fa3a79c2c859c5daa6e562a73328ce422b322357e666cf084"} Dec 03 16:46:14 crc kubenswrapper[4998]: I1203 16:46:14.577857 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kg9v4" event={"ID":"beac9431-5b94-4cec-b29c-7c01f9de8e59","Type":"ContainerStarted","Data":"89169f290c187a0e89351478ab46129aa757661620f099950b4916a2ae2ecd72"} Dec 03 16:46:14 crc kubenswrapper[4998]: I1203 16:46:14.603141 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kg9v4" podStartSLOduration=2.20160988 podStartE2EDuration="3.603122944s" podCreationTimestamp="2025-12-03 16:46:11 +0000 UTC" firstStartedPulling="2025-12-03 16:46:12.555481121 +0000 UTC m=+2551.167181344" lastFinishedPulling="2025-12-03 16:46:13.956994175 +0000 UTC m=+2552.568694408" observedRunningTime="2025-12-03 16:46:14.595793254 +0000 UTC m=+2553.207493497" watchObservedRunningTime="2025-12-03 16:46:14.603122944 +0000 UTC m=+2553.214823157" Dec 03 16:46:21 crc kubenswrapper[4998]: I1203 16:46:21.618212 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kg9v4" Dec 03 16:46:21 crc kubenswrapper[4998]: I1203 16:46:21.618862 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kg9v4" Dec 03 16:46:21 crc kubenswrapper[4998]: I1203 16:46:21.692464 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kg9v4" Dec 03 16:46:21 crc kubenswrapper[4998]: I1203 16:46:21.749526 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kg9v4" Dec 03 16:46:21 crc kubenswrapper[4998]: I1203 16:46:21.933250 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kg9v4"] Dec 03 16:46:23 crc kubenswrapper[4998]: I1203 16:46:23.664340 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kg9v4" podUID="beac9431-5b94-4cec-b29c-7c01f9de8e59" containerName="registry-server" containerID="cri-o://89169f290c187a0e89351478ab46129aa757661620f099950b4916a2ae2ecd72" gracePeriod=2 Dec 03 16:46:24 crc kubenswrapper[4998]: I1203 16:46:24.682308 4998 generic.go:334] "Generic (PLEG): container finished" podID="beac9431-5b94-4cec-b29c-7c01f9de8e59" containerID="89169f290c187a0e89351478ab46129aa757661620f099950b4916a2ae2ecd72" exitCode=0 Dec 03 16:46:24 crc kubenswrapper[4998]: I1203 16:46:24.682404 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kg9v4" event={"ID":"beac9431-5b94-4cec-b29c-7c01f9de8e59","Type":"ContainerDied","Data":"89169f290c187a0e89351478ab46129aa757661620f099950b4916a2ae2ecd72"} Dec 03 16:46:24 crc kubenswrapper[4998]: I1203 16:46:24.822832 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kg9v4" Dec 03 16:46:24 crc kubenswrapper[4998]: I1203 16:46:24.934040 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/beac9431-5b94-4cec-b29c-7c01f9de8e59-catalog-content\") pod \"beac9431-5b94-4cec-b29c-7c01f9de8e59\" (UID: \"beac9431-5b94-4cec-b29c-7c01f9de8e59\") " Dec 03 16:46:24 crc kubenswrapper[4998]: I1203 16:46:24.934189 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gb4xz\" (UniqueName: \"kubernetes.io/projected/beac9431-5b94-4cec-b29c-7c01f9de8e59-kube-api-access-gb4xz\") pod \"beac9431-5b94-4cec-b29c-7c01f9de8e59\" (UID: \"beac9431-5b94-4cec-b29c-7c01f9de8e59\") " Dec 03 16:46:24 crc kubenswrapper[4998]: I1203 16:46:24.935250 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/beac9431-5b94-4cec-b29c-7c01f9de8e59-utilities\") pod \"beac9431-5b94-4cec-b29c-7c01f9de8e59\" (UID: \"beac9431-5b94-4cec-b29c-7c01f9de8e59\") " Dec 03 16:46:24 crc kubenswrapper[4998]: I1203 16:46:24.936033 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/beac9431-5b94-4cec-b29c-7c01f9de8e59-utilities" (OuterVolumeSpecName: "utilities") pod "beac9431-5b94-4cec-b29c-7c01f9de8e59" (UID: "beac9431-5b94-4cec-b29c-7c01f9de8e59"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:46:24 crc kubenswrapper[4998]: I1203 16:46:24.939111 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/beac9431-5b94-4cec-b29c-7c01f9de8e59-kube-api-access-gb4xz" (OuterVolumeSpecName: "kube-api-access-gb4xz") pod "beac9431-5b94-4cec-b29c-7c01f9de8e59" (UID: "beac9431-5b94-4cec-b29c-7c01f9de8e59"). InnerVolumeSpecName "kube-api-access-gb4xz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:46:24 crc kubenswrapper[4998]: I1203 16:46:24.950389 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/beac9431-5b94-4cec-b29c-7c01f9de8e59-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "beac9431-5b94-4cec-b29c-7c01f9de8e59" (UID: "beac9431-5b94-4cec-b29c-7c01f9de8e59"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:46:25 crc kubenswrapper[4998]: I1203 16:46:25.037147 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gb4xz\" (UniqueName: \"kubernetes.io/projected/beac9431-5b94-4cec-b29c-7c01f9de8e59-kube-api-access-gb4xz\") on node \"crc\" DevicePath \"\"" Dec 03 16:46:25 crc kubenswrapper[4998]: I1203 16:46:25.037182 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/beac9431-5b94-4cec-b29c-7c01f9de8e59-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:46:25 crc kubenswrapper[4998]: I1203 16:46:25.037191 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/beac9431-5b94-4cec-b29c-7c01f9de8e59-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:46:25 crc kubenswrapper[4998]: I1203 16:46:25.715512 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kg9v4" event={"ID":"beac9431-5b94-4cec-b29c-7c01f9de8e59","Type":"ContainerDied","Data":"bb8ddb712cf32426a8583ce57da5650a6e099745b1e72f70241bd71f1e4a6823"} Dec 03 16:46:25 crc kubenswrapper[4998]: I1203 16:46:25.715612 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kg9v4" Dec 03 16:46:25 crc kubenswrapper[4998]: I1203 16:46:25.716055 4998 scope.go:117] "RemoveContainer" containerID="89169f290c187a0e89351478ab46129aa757661620f099950b4916a2ae2ecd72" Dec 03 16:46:25 crc kubenswrapper[4998]: I1203 16:46:25.749277 4998 scope.go:117] "RemoveContainer" containerID="c58cce204fa47a5fa3a79c2c859c5daa6e562a73328ce422b322357e666cf084" Dec 03 16:46:25 crc kubenswrapper[4998]: I1203 16:46:25.761697 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kg9v4"] Dec 03 16:46:25 crc kubenswrapper[4998]: I1203 16:46:25.770798 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kg9v4"] Dec 03 16:46:25 crc kubenswrapper[4998]: I1203 16:46:25.775343 4998 scope.go:117] "RemoveContainer" containerID="69d73dbecd2728b9454a688e1b923605e3d866412d7365a25e9ae4d45251c8d4" Dec 03 16:46:27 crc kubenswrapper[4998]: I1203 16:46:27.692130 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="beac9431-5b94-4cec-b29c-7c01f9de8e59" path="/var/lib/kubelet/pods/beac9431-5b94-4cec-b29c-7c01f9de8e59/volumes" Dec 03 16:46:52 crc kubenswrapper[4998]: I1203 16:46:52.594363 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rdl8w"] Dec 03 16:46:52 crc kubenswrapper[4998]: E1203 16:46:52.595219 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beac9431-5b94-4cec-b29c-7c01f9de8e59" containerName="extract-content" Dec 03 16:46:52 crc kubenswrapper[4998]: I1203 16:46:52.595231 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="beac9431-5b94-4cec-b29c-7c01f9de8e59" containerName="extract-content" Dec 03 16:46:52 crc kubenswrapper[4998]: E1203 16:46:52.595252 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beac9431-5b94-4cec-b29c-7c01f9de8e59" containerName="registry-server" Dec 03 16:46:52 crc kubenswrapper[4998]: I1203 16:46:52.595258 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="beac9431-5b94-4cec-b29c-7c01f9de8e59" containerName="registry-server" Dec 03 16:46:52 crc kubenswrapper[4998]: E1203 16:46:52.595279 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beac9431-5b94-4cec-b29c-7c01f9de8e59" containerName="extract-utilities" Dec 03 16:46:52 crc kubenswrapper[4998]: I1203 16:46:52.595284 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="beac9431-5b94-4cec-b29c-7c01f9de8e59" containerName="extract-utilities" Dec 03 16:46:52 crc kubenswrapper[4998]: I1203 16:46:52.595477 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="beac9431-5b94-4cec-b29c-7c01f9de8e59" containerName="registry-server" Dec 03 16:46:52 crc kubenswrapper[4998]: I1203 16:46:52.597004 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rdl8w" Dec 03 16:46:52 crc kubenswrapper[4998]: I1203 16:46:52.606089 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rdl8w"] Dec 03 16:46:52 crc kubenswrapper[4998]: I1203 16:46:52.643976 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f825c8c-4b4d-466f-92d8-8e50a3c4a831-catalog-content\") pod \"redhat-operators-rdl8w\" (UID: \"1f825c8c-4b4d-466f-92d8-8e50a3c4a831\") " pod="openshift-marketplace/redhat-operators-rdl8w" Dec 03 16:46:52 crc kubenswrapper[4998]: I1203 16:46:52.644055 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f825c8c-4b4d-466f-92d8-8e50a3c4a831-utilities\") pod \"redhat-operators-rdl8w\" (UID: \"1f825c8c-4b4d-466f-92d8-8e50a3c4a831\") " pod="openshift-marketplace/redhat-operators-rdl8w" Dec 03 16:46:52 crc kubenswrapper[4998]: I1203 16:46:52.644227 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6qtf\" (UniqueName: \"kubernetes.io/projected/1f825c8c-4b4d-466f-92d8-8e50a3c4a831-kube-api-access-x6qtf\") pod \"redhat-operators-rdl8w\" (UID: \"1f825c8c-4b4d-466f-92d8-8e50a3c4a831\") " pod="openshift-marketplace/redhat-operators-rdl8w" Dec 03 16:46:52 crc kubenswrapper[4998]: I1203 16:46:52.746472 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f825c8c-4b4d-466f-92d8-8e50a3c4a831-catalog-content\") pod \"redhat-operators-rdl8w\" (UID: \"1f825c8c-4b4d-466f-92d8-8e50a3c4a831\") " pod="openshift-marketplace/redhat-operators-rdl8w" Dec 03 16:46:52 crc kubenswrapper[4998]: I1203 16:46:52.746526 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f825c8c-4b4d-466f-92d8-8e50a3c4a831-utilities\") pod \"redhat-operators-rdl8w\" (UID: \"1f825c8c-4b4d-466f-92d8-8e50a3c4a831\") " pod="openshift-marketplace/redhat-operators-rdl8w" Dec 03 16:46:52 crc kubenswrapper[4998]: I1203 16:46:52.746570 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6qtf\" (UniqueName: \"kubernetes.io/projected/1f825c8c-4b4d-466f-92d8-8e50a3c4a831-kube-api-access-x6qtf\") pod \"redhat-operators-rdl8w\" (UID: \"1f825c8c-4b4d-466f-92d8-8e50a3c4a831\") " pod="openshift-marketplace/redhat-operators-rdl8w" Dec 03 16:46:52 crc kubenswrapper[4998]: I1203 16:46:52.747256 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f825c8c-4b4d-466f-92d8-8e50a3c4a831-catalog-content\") pod \"redhat-operators-rdl8w\" (UID: \"1f825c8c-4b4d-466f-92d8-8e50a3c4a831\") " pod="openshift-marketplace/redhat-operators-rdl8w" Dec 03 16:46:52 crc kubenswrapper[4998]: I1203 16:46:52.747344 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f825c8c-4b4d-466f-92d8-8e50a3c4a831-utilities\") pod \"redhat-operators-rdl8w\" (UID: \"1f825c8c-4b4d-466f-92d8-8e50a3c4a831\") " pod="openshift-marketplace/redhat-operators-rdl8w" Dec 03 16:46:52 crc kubenswrapper[4998]: I1203 16:46:52.767602 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6qtf\" (UniqueName: \"kubernetes.io/projected/1f825c8c-4b4d-466f-92d8-8e50a3c4a831-kube-api-access-x6qtf\") pod \"redhat-operators-rdl8w\" (UID: \"1f825c8c-4b4d-466f-92d8-8e50a3c4a831\") " pod="openshift-marketplace/redhat-operators-rdl8w" Dec 03 16:46:52 crc kubenswrapper[4998]: I1203 16:46:52.928711 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rdl8w" Dec 03 16:46:54 crc kubenswrapper[4998]: I1203 16:46:54.470905 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rdl8w"] Dec 03 16:46:54 crc kubenswrapper[4998]: W1203 16:46:54.497997 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f825c8c_4b4d_466f_92d8_8e50a3c4a831.slice/crio-1f179ff363deef15b096275d97ecded3283aabb29773497d476b868bc8452f03 WatchSource:0}: Error finding container 1f179ff363deef15b096275d97ecded3283aabb29773497d476b868bc8452f03: Status 404 returned error can't find the container with id 1f179ff363deef15b096275d97ecded3283aabb29773497d476b868bc8452f03 Dec 03 16:46:55 crc kubenswrapper[4998]: I1203 16:46:55.017247 4998 generic.go:334] "Generic (PLEG): container finished" podID="1f825c8c-4b4d-466f-92d8-8e50a3c4a831" containerID="70a5c82d3dbc3bab0169c8327e65db9de3c9a317d138defddb3f54a552dc0c2e" exitCode=0 Dec 03 16:46:55 crc kubenswrapper[4998]: I1203 16:46:55.017390 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdl8w" event={"ID":"1f825c8c-4b4d-466f-92d8-8e50a3c4a831","Type":"ContainerDied","Data":"70a5c82d3dbc3bab0169c8327e65db9de3c9a317d138defddb3f54a552dc0c2e"} Dec 03 16:46:55 crc kubenswrapper[4998]: I1203 16:46:55.017972 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdl8w" event={"ID":"1f825c8c-4b4d-466f-92d8-8e50a3c4a831","Type":"ContainerStarted","Data":"1f179ff363deef15b096275d97ecded3283aabb29773497d476b868bc8452f03"} Dec 03 16:46:55 crc kubenswrapper[4998]: I1203 16:46:55.019331 4998 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 16:46:56 crc kubenswrapper[4998]: I1203 16:46:56.030214 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdl8w" event={"ID":"1f825c8c-4b4d-466f-92d8-8e50a3c4a831","Type":"ContainerStarted","Data":"feb1cc1be3ef97ee8a0562519f4a260ba031e17e0a8286b202ad0f19673aa7cb"} Dec 03 16:46:59 crc kubenswrapper[4998]: I1203 16:46:59.061288 4998 generic.go:334] "Generic (PLEG): container finished" podID="1f825c8c-4b4d-466f-92d8-8e50a3c4a831" containerID="feb1cc1be3ef97ee8a0562519f4a260ba031e17e0a8286b202ad0f19673aa7cb" exitCode=0 Dec 03 16:46:59 crc kubenswrapper[4998]: I1203 16:46:59.061363 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdl8w" event={"ID":"1f825c8c-4b4d-466f-92d8-8e50a3c4a831","Type":"ContainerDied","Data":"feb1cc1be3ef97ee8a0562519f4a260ba031e17e0a8286b202ad0f19673aa7cb"} Dec 03 16:47:01 crc kubenswrapper[4998]: I1203 16:47:01.086126 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdl8w" event={"ID":"1f825c8c-4b4d-466f-92d8-8e50a3c4a831","Type":"ContainerStarted","Data":"e9f0649982ccc35ad40d29b864c760a7383130bc7c5f05beda0193e56d983ae7"} Dec 03 16:47:01 crc kubenswrapper[4998]: I1203 16:47:01.107646 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rdl8w" podStartSLOduration=3.298957584 podStartE2EDuration="9.107626487s" podCreationTimestamp="2025-12-03 16:46:52 +0000 UTC" firstStartedPulling="2025-12-03 16:46:55.019117593 +0000 UTC m=+2593.630817816" lastFinishedPulling="2025-12-03 16:47:00.827786476 +0000 UTC m=+2599.439486719" observedRunningTime="2025-12-03 16:47:01.104184123 +0000 UTC m=+2599.715884366" watchObservedRunningTime="2025-12-03 16:47:01.107626487 +0000 UTC m=+2599.719326710" Dec 03 16:47:02 crc kubenswrapper[4998]: I1203 16:47:02.929012 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rdl8w" Dec 03 16:47:02 crc kubenswrapper[4998]: I1203 16:47:02.929367 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rdl8w" Dec 03 16:47:03 crc kubenswrapper[4998]: I1203 16:47:03.977360 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rdl8w" podUID="1f825c8c-4b4d-466f-92d8-8e50a3c4a831" containerName="registry-server" probeResult="failure" output=< Dec 03 16:47:03 crc kubenswrapper[4998]: timeout: failed to connect service ":50051" within 1s Dec 03 16:47:03 crc kubenswrapper[4998]: > Dec 03 16:47:12 crc kubenswrapper[4998]: I1203 16:47:12.994917 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rdl8w" Dec 03 16:47:13 crc kubenswrapper[4998]: I1203 16:47:13.065011 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rdl8w" Dec 03 16:47:13 crc kubenswrapper[4998]: I1203 16:47:13.245627 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rdl8w"] Dec 03 16:47:14 crc kubenswrapper[4998]: I1203 16:47:14.214491 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rdl8w" podUID="1f825c8c-4b4d-466f-92d8-8e50a3c4a831" containerName="registry-server" containerID="cri-o://e9f0649982ccc35ad40d29b864c760a7383130bc7c5f05beda0193e56d983ae7" gracePeriod=2 Dec 03 16:47:14 crc kubenswrapper[4998]: I1203 16:47:14.758018 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rdl8w" Dec 03 16:47:14 crc kubenswrapper[4998]: I1203 16:47:14.822578 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f825c8c-4b4d-466f-92d8-8e50a3c4a831-utilities\") pod \"1f825c8c-4b4d-466f-92d8-8e50a3c4a831\" (UID: \"1f825c8c-4b4d-466f-92d8-8e50a3c4a831\") " Dec 03 16:47:14 crc kubenswrapper[4998]: I1203 16:47:14.822925 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f825c8c-4b4d-466f-92d8-8e50a3c4a831-catalog-content\") pod \"1f825c8c-4b4d-466f-92d8-8e50a3c4a831\" (UID: \"1f825c8c-4b4d-466f-92d8-8e50a3c4a831\") " Dec 03 16:47:14 crc kubenswrapper[4998]: I1203 16:47:14.822961 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6qtf\" (UniqueName: \"kubernetes.io/projected/1f825c8c-4b4d-466f-92d8-8e50a3c4a831-kube-api-access-x6qtf\") pod \"1f825c8c-4b4d-466f-92d8-8e50a3c4a831\" (UID: \"1f825c8c-4b4d-466f-92d8-8e50a3c4a831\") " Dec 03 16:47:14 crc kubenswrapper[4998]: I1203 16:47:14.824254 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f825c8c-4b4d-466f-92d8-8e50a3c4a831-utilities" (OuterVolumeSpecName: "utilities") pod "1f825c8c-4b4d-466f-92d8-8e50a3c4a831" (UID: "1f825c8c-4b4d-466f-92d8-8e50a3c4a831"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:47:14 crc kubenswrapper[4998]: I1203 16:47:14.829196 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f825c8c-4b4d-466f-92d8-8e50a3c4a831-kube-api-access-x6qtf" (OuterVolumeSpecName: "kube-api-access-x6qtf") pod "1f825c8c-4b4d-466f-92d8-8e50a3c4a831" (UID: "1f825c8c-4b4d-466f-92d8-8e50a3c4a831"). InnerVolumeSpecName "kube-api-access-x6qtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:47:14 crc kubenswrapper[4998]: I1203 16:47:14.925417 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f825c8c-4b4d-466f-92d8-8e50a3c4a831-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:47:14 crc kubenswrapper[4998]: I1203 16:47:14.925475 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6qtf\" (UniqueName: \"kubernetes.io/projected/1f825c8c-4b4d-466f-92d8-8e50a3c4a831-kube-api-access-x6qtf\") on node \"crc\" DevicePath \"\"" Dec 03 16:47:14 crc kubenswrapper[4998]: I1203 16:47:14.945043 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f825c8c-4b4d-466f-92d8-8e50a3c4a831-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1f825c8c-4b4d-466f-92d8-8e50a3c4a831" (UID: "1f825c8c-4b4d-466f-92d8-8e50a3c4a831"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:47:15 crc kubenswrapper[4998]: I1203 16:47:15.027834 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f825c8c-4b4d-466f-92d8-8e50a3c4a831-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:47:15 crc kubenswrapper[4998]: I1203 16:47:15.227008 4998 generic.go:334] "Generic (PLEG): container finished" podID="1f825c8c-4b4d-466f-92d8-8e50a3c4a831" containerID="e9f0649982ccc35ad40d29b864c760a7383130bc7c5f05beda0193e56d983ae7" exitCode=0 Dec 03 16:47:15 crc kubenswrapper[4998]: I1203 16:47:15.227059 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdl8w" event={"ID":"1f825c8c-4b4d-466f-92d8-8e50a3c4a831","Type":"ContainerDied","Data":"e9f0649982ccc35ad40d29b864c760a7383130bc7c5f05beda0193e56d983ae7"} Dec 03 16:47:15 crc kubenswrapper[4998]: I1203 16:47:15.227095 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdl8w" event={"ID":"1f825c8c-4b4d-466f-92d8-8e50a3c4a831","Type":"ContainerDied","Data":"1f179ff363deef15b096275d97ecded3283aabb29773497d476b868bc8452f03"} Dec 03 16:47:15 crc kubenswrapper[4998]: I1203 16:47:15.227116 4998 scope.go:117] "RemoveContainer" containerID="e9f0649982ccc35ad40d29b864c760a7383130bc7c5f05beda0193e56d983ae7" Dec 03 16:47:15 crc kubenswrapper[4998]: I1203 16:47:15.227154 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rdl8w" Dec 03 16:47:15 crc kubenswrapper[4998]: I1203 16:47:15.250424 4998 scope.go:117] "RemoveContainer" containerID="feb1cc1be3ef97ee8a0562519f4a260ba031e17e0a8286b202ad0f19673aa7cb" Dec 03 16:47:15 crc kubenswrapper[4998]: I1203 16:47:15.265080 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rdl8w"] Dec 03 16:47:15 crc kubenswrapper[4998]: I1203 16:47:15.274724 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rdl8w"] Dec 03 16:47:15 crc kubenswrapper[4998]: I1203 16:47:15.298653 4998 scope.go:117] "RemoveContainer" containerID="70a5c82d3dbc3bab0169c8327e65db9de3c9a317d138defddb3f54a552dc0c2e" Dec 03 16:47:15 crc kubenswrapper[4998]: I1203 16:47:15.323858 4998 scope.go:117] "RemoveContainer" containerID="e9f0649982ccc35ad40d29b864c760a7383130bc7c5f05beda0193e56d983ae7" Dec 03 16:47:15 crc kubenswrapper[4998]: E1203 16:47:15.324297 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9f0649982ccc35ad40d29b864c760a7383130bc7c5f05beda0193e56d983ae7\": container with ID starting with e9f0649982ccc35ad40d29b864c760a7383130bc7c5f05beda0193e56d983ae7 not found: ID does not exist" containerID="e9f0649982ccc35ad40d29b864c760a7383130bc7c5f05beda0193e56d983ae7" Dec 03 16:47:15 crc kubenswrapper[4998]: I1203 16:47:15.324335 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9f0649982ccc35ad40d29b864c760a7383130bc7c5f05beda0193e56d983ae7"} err="failed to get container status \"e9f0649982ccc35ad40d29b864c760a7383130bc7c5f05beda0193e56d983ae7\": rpc error: code = NotFound desc = could not find container \"e9f0649982ccc35ad40d29b864c760a7383130bc7c5f05beda0193e56d983ae7\": container with ID starting with e9f0649982ccc35ad40d29b864c760a7383130bc7c5f05beda0193e56d983ae7 not found: ID does not exist" Dec 03 16:47:15 crc kubenswrapper[4998]: I1203 16:47:15.324396 4998 scope.go:117] "RemoveContainer" containerID="feb1cc1be3ef97ee8a0562519f4a260ba031e17e0a8286b202ad0f19673aa7cb" Dec 03 16:47:15 crc kubenswrapper[4998]: E1203 16:47:15.324644 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"feb1cc1be3ef97ee8a0562519f4a260ba031e17e0a8286b202ad0f19673aa7cb\": container with ID starting with feb1cc1be3ef97ee8a0562519f4a260ba031e17e0a8286b202ad0f19673aa7cb not found: ID does not exist" containerID="feb1cc1be3ef97ee8a0562519f4a260ba031e17e0a8286b202ad0f19673aa7cb" Dec 03 16:47:15 crc kubenswrapper[4998]: I1203 16:47:15.324676 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"feb1cc1be3ef97ee8a0562519f4a260ba031e17e0a8286b202ad0f19673aa7cb"} err="failed to get container status \"feb1cc1be3ef97ee8a0562519f4a260ba031e17e0a8286b202ad0f19673aa7cb\": rpc error: code = NotFound desc = could not find container \"feb1cc1be3ef97ee8a0562519f4a260ba031e17e0a8286b202ad0f19673aa7cb\": container with ID starting with feb1cc1be3ef97ee8a0562519f4a260ba031e17e0a8286b202ad0f19673aa7cb not found: ID does not exist" Dec 03 16:47:15 crc kubenswrapper[4998]: I1203 16:47:15.324694 4998 scope.go:117] "RemoveContainer" containerID="70a5c82d3dbc3bab0169c8327e65db9de3c9a317d138defddb3f54a552dc0c2e" Dec 03 16:47:15 crc kubenswrapper[4998]: E1203 16:47:15.325039 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70a5c82d3dbc3bab0169c8327e65db9de3c9a317d138defddb3f54a552dc0c2e\": container with ID starting with 70a5c82d3dbc3bab0169c8327e65db9de3c9a317d138defddb3f54a552dc0c2e not found: ID does not exist" containerID="70a5c82d3dbc3bab0169c8327e65db9de3c9a317d138defddb3f54a552dc0c2e" Dec 03 16:47:15 crc kubenswrapper[4998]: I1203 16:47:15.325079 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70a5c82d3dbc3bab0169c8327e65db9de3c9a317d138defddb3f54a552dc0c2e"} err="failed to get container status \"70a5c82d3dbc3bab0169c8327e65db9de3c9a317d138defddb3f54a552dc0c2e\": rpc error: code = NotFound desc = could not find container \"70a5c82d3dbc3bab0169c8327e65db9de3c9a317d138defddb3f54a552dc0c2e\": container with ID starting with 70a5c82d3dbc3bab0169c8327e65db9de3c9a317d138defddb3f54a552dc0c2e not found: ID does not exist" Dec 03 16:47:15 crc kubenswrapper[4998]: I1203 16:47:15.691308 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f825c8c-4b4d-466f-92d8-8e50a3c4a831" path="/var/lib/kubelet/pods/1f825c8c-4b4d-466f-92d8-8e50a3c4a831/volumes" Dec 03 16:47:27 crc kubenswrapper[4998]: I1203 16:47:27.111732 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:47:27 crc kubenswrapper[4998]: I1203 16:47:27.112496 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:47:57 crc kubenswrapper[4998]: I1203 16:47:57.111083 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:47:57 crc kubenswrapper[4998]: I1203 16:47:57.111599 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:48:26 crc kubenswrapper[4998]: I1203 16:48:26.956203 4998 generic.go:334] "Generic (PLEG): container finished" podID="d23ecf9c-0347-4a37-ae82-c7b24d5b9575" containerID="d632ab7fddf239b15629f7e135d77abd3b1758ee543f3caead3ec023967fd3be" exitCode=0 Dec 03 16:48:26 crc kubenswrapper[4998]: I1203 16:48:26.956795 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q" event={"ID":"d23ecf9c-0347-4a37-ae82-c7b24d5b9575","Type":"ContainerDied","Data":"d632ab7fddf239b15629f7e135d77abd3b1758ee543f3caead3ec023967fd3be"} Dec 03 16:48:27 crc kubenswrapper[4998]: I1203 16:48:27.111076 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:48:27 crc kubenswrapper[4998]: I1203 16:48:27.111171 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:48:27 crc kubenswrapper[4998]: I1203 16:48:27.111235 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:48:27 crc kubenswrapper[4998]: I1203 16:48:27.112490 4998 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"65e608d68983552183ea47364a0c23e120fed4872f3ce284dfdd07db6b645cc3"} pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 16:48:27 crc kubenswrapper[4998]: I1203 16:48:27.112601 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" containerID="cri-o://65e608d68983552183ea47364a0c23e120fed4872f3ce284dfdd07db6b645cc3" gracePeriod=600 Dec 03 16:48:27 crc kubenswrapper[4998]: I1203 16:48:27.968816 4998 generic.go:334] "Generic (PLEG): container finished" podID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerID="65e608d68983552183ea47364a0c23e120fed4872f3ce284dfdd07db6b645cc3" exitCode=0 Dec 03 16:48:27 crc kubenswrapper[4998]: I1203 16:48:27.968857 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerDied","Data":"65e608d68983552183ea47364a0c23e120fed4872f3ce284dfdd07db6b645cc3"} Dec 03 16:48:27 crc kubenswrapper[4998]: I1203 16:48:27.969459 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096"} Dec 03 16:48:27 crc kubenswrapper[4998]: I1203 16:48:27.969508 4998 scope.go:117] "RemoveContainer" containerID="8ca88cecdfa82d1e9775baa178b3ca886ee9ff9fdcba6662420b296663e9e155" Dec 03 16:48:28 crc kubenswrapper[4998]: I1203 16:48:28.432808 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q" Dec 03 16:48:28 crc kubenswrapper[4998]: I1203 16:48:28.606700 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-libvirt-combined-ca-bundle\") pod \"d23ecf9c-0347-4a37-ae82-c7b24d5b9575\" (UID: \"d23ecf9c-0347-4a37-ae82-c7b24d5b9575\") " Dec 03 16:48:28 crc kubenswrapper[4998]: I1203 16:48:28.606896 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-ssh-key\") pod \"d23ecf9c-0347-4a37-ae82-c7b24d5b9575\" (UID: \"d23ecf9c-0347-4a37-ae82-c7b24d5b9575\") " Dec 03 16:48:28 crc kubenswrapper[4998]: I1203 16:48:28.606977 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fp2n\" (UniqueName: \"kubernetes.io/projected/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-kube-api-access-9fp2n\") pod \"d23ecf9c-0347-4a37-ae82-c7b24d5b9575\" (UID: \"d23ecf9c-0347-4a37-ae82-c7b24d5b9575\") " Dec 03 16:48:28 crc kubenswrapper[4998]: I1203 16:48:28.607003 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-libvirt-secret-0\") pod \"d23ecf9c-0347-4a37-ae82-c7b24d5b9575\" (UID: \"d23ecf9c-0347-4a37-ae82-c7b24d5b9575\") " Dec 03 16:48:28 crc kubenswrapper[4998]: I1203 16:48:28.607057 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-inventory\") pod \"d23ecf9c-0347-4a37-ae82-c7b24d5b9575\" (UID: \"d23ecf9c-0347-4a37-ae82-c7b24d5b9575\") " Dec 03 16:48:28 crc kubenswrapper[4998]: I1203 16:48:28.616625 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-kube-api-access-9fp2n" (OuterVolumeSpecName: "kube-api-access-9fp2n") pod "d23ecf9c-0347-4a37-ae82-c7b24d5b9575" (UID: "d23ecf9c-0347-4a37-ae82-c7b24d5b9575"). InnerVolumeSpecName "kube-api-access-9fp2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:48:28 crc kubenswrapper[4998]: I1203 16:48:28.617728 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "d23ecf9c-0347-4a37-ae82-c7b24d5b9575" (UID: "d23ecf9c-0347-4a37-ae82-c7b24d5b9575"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:48:28 crc kubenswrapper[4998]: I1203 16:48:28.650267 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-inventory" (OuterVolumeSpecName: "inventory") pod "d23ecf9c-0347-4a37-ae82-c7b24d5b9575" (UID: "d23ecf9c-0347-4a37-ae82-c7b24d5b9575"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:48:28 crc kubenswrapper[4998]: I1203 16:48:28.656518 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "d23ecf9c-0347-4a37-ae82-c7b24d5b9575" (UID: "d23ecf9c-0347-4a37-ae82-c7b24d5b9575"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:48:28 crc kubenswrapper[4998]: I1203 16:48:28.658946 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d23ecf9c-0347-4a37-ae82-c7b24d5b9575" (UID: "d23ecf9c-0347-4a37-ae82-c7b24d5b9575"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:48:28 crc kubenswrapper[4998]: I1203 16:48:28.713488 4998 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 16:48:28 crc kubenswrapper[4998]: I1203 16:48:28.713774 4998 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:48:28 crc kubenswrapper[4998]: I1203 16:48:28.713928 4998 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 16:48:28 crc kubenswrapper[4998]: I1203 16:48:28.714006 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fp2n\" (UniqueName: \"kubernetes.io/projected/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-kube-api-access-9fp2n\") on node \"crc\" DevicePath \"\"" Dec 03 16:48:28 crc kubenswrapper[4998]: I1203 16:48:28.714089 4998 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/d23ecf9c-0347-4a37-ae82-c7b24d5b9575-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 03 16:48:28 crc kubenswrapper[4998]: I1203 16:48:28.982973 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q" event={"ID":"d23ecf9c-0347-4a37-ae82-c7b24d5b9575","Type":"ContainerDied","Data":"53be1987a9081f6cf510961134fb3aa23cb44e7858a06343d7a0bfcbe1a127ed"} Dec 03 16:48:28 crc kubenswrapper[4998]: I1203 16:48:28.983021 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53be1987a9081f6cf510961134fb3aa23cb44e7858a06343d7a0bfcbe1a127ed" Dec 03 16:48:28 crc kubenswrapper[4998]: I1203 16:48:28.983103 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.135573 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l"] Dec 03 16:48:29 crc kubenswrapper[4998]: E1203 16:48:29.135981 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d23ecf9c-0347-4a37-ae82-c7b24d5b9575" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.135998 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="d23ecf9c-0347-4a37-ae82-c7b24d5b9575" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 16:48:29 crc kubenswrapper[4998]: E1203 16:48:29.136012 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f825c8c-4b4d-466f-92d8-8e50a3c4a831" containerName="extract-utilities" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.136020 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f825c8c-4b4d-466f-92d8-8e50a3c4a831" containerName="extract-utilities" Dec 03 16:48:29 crc kubenswrapper[4998]: E1203 16:48:29.136039 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f825c8c-4b4d-466f-92d8-8e50a3c4a831" containerName="registry-server" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.136046 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f825c8c-4b4d-466f-92d8-8e50a3c4a831" containerName="registry-server" Dec 03 16:48:29 crc kubenswrapper[4998]: E1203 16:48:29.136063 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f825c8c-4b4d-466f-92d8-8e50a3c4a831" containerName="extract-content" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.136069 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f825c8c-4b4d-466f-92d8-8e50a3c4a831" containerName="extract-content" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.136251 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="d23ecf9c-0347-4a37-ae82-c7b24d5b9575" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.136262 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f825c8c-4b4d-466f-92d8-8e50a3c4a831" containerName="registry-server" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.136959 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.144884 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.144906 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.144924 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.145050 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.145111 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.145160 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.145320 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kcj9x" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.157166 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l"] Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.160947 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.161020 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.161046 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4hfp\" (UniqueName: \"kubernetes.io/projected/5af60798-c3fe-4b82-9732-4cfd4c442549-kube-api-access-n4hfp\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.161077 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.161099 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.161175 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.161274 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.161299 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.161318 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.263048 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.263088 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.263166 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.263248 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.263268 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.263285 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.263322 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.263349 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.263365 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4hfp\" (UniqueName: \"kubernetes.io/projected/5af60798-c3fe-4b82-9732-4cfd4c442549-kube-api-access-n4hfp\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.265034 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.268614 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.269032 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.269153 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.270492 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.274673 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.282463 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.282737 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.283967 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4hfp\" (UniqueName: \"kubernetes.io/projected/5af60798-c3fe-4b82-9732-4cfd4c442549-kube-api-access-n4hfp\") pod \"nova-edpm-deployment-openstack-edpm-ipam-w6l7l\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:29 crc kubenswrapper[4998]: I1203 16:48:29.453179 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:48:30 crc kubenswrapper[4998]: I1203 16:48:30.010732 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l"] Dec 03 16:48:30 crc kubenswrapper[4998]: W1203 16:48:30.012695 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5af60798_c3fe_4b82_9732_4cfd4c442549.slice/crio-a49cdd635df653fab64a83df37257dd4cb85c54ecc5357bd6dcc91b73265238b WatchSource:0}: Error finding container a49cdd635df653fab64a83df37257dd4cb85c54ecc5357bd6dcc91b73265238b: Status 404 returned error can't find the container with id a49cdd635df653fab64a83df37257dd4cb85c54ecc5357bd6dcc91b73265238b Dec 03 16:48:31 crc kubenswrapper[4998]: I1203 16:48:31.010048 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" event={"ID":"5af60798-c3fe-4b82-9732-4cfd4c442549","Type":"ContainerStarted","Data":"a49cdd635df653fab64a83df37257dd4cb85c54ecc5357bd6dcc91b73265238b"} Dec 03 16:48:32 crc kubenswrapper[4998]: I1203 16:48:32.023369 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" event={"ID":"5af60798-c3fe-4b82-9732-4cfd4c442549","Type":"ContainerStarted","Data":"d35a6171cb2a45a70783e1f5f6031a2297c93bc0b6e3be537740b47b3b0daa83"} Dec 03 16:48:33 crc kubenswrapper[4998]: I1203 16:48:33.053127 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" podStartSLOduration=2.58387591 podStartE2EDuration="4.053111445s" podCreationTimestamp="2025-12-03 16:48:29 +0000 UTC" firstStartedPulling="2025-12-03 16:48:30.018790666 +0000 UTC m=+2688.630490889" lastFinishedPulling="2025-12-03 16:48:31.488026201 +0000 UTC m=+2690.099726424" observedRunningTime="2025-12-03 16:48:33.047322583 +0000 UTC m=+2691.659022816" watchObservedRunningTime="2025-12-03 16:48:33.053111445 +0000 UTC m=+2691.664811668" Dec 03 16:49:35 crc kubenswrapper[4998]: I1203 16:49:35.368601 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r9dc9"] Dec 03 16:49:35 crc kubenswrapper[4998]: I1203 16:49:35.373386 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r9dc9" Dec 03 16:49:35 crc kubenswrapper[4998]: I1203 16:49:35.418247 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r9dc9"] Dec 03 16:49:35 crc kubenswrapper[4998]: I1203 16:49:35.434156 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9759ee01-8377-4d37-b262-dd591a841bd2-utilities\") pod \"certified-operators-r9dc9\" (UID: \"9759ee01-8377-4d37-b262-dd591a841bd2\") " pod="openshift-marketplace/certified-operators-r9dc9" Dec 03 16:49:35 crc kubenswrapper[4998]: I1203 16:49:35.434278 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9759ee01-8377-4d37-b262-dd591a841bd2-catalog-content\") pod \"certified-operators-r9dc9\" (UID: \"9759ee01-8377-4d37-b262-dd591a841bd2\") " pod="openshift-marketplace/certified-operators-r9dc9" Dec 03 16:49:35 crc kubenswrapper[4998]: I1203 16:49:35.434423 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxpsj\" (UniqueName: \"kubernetes.io/projected/9759ee01-8377-4d37-b262-dd591a841bd2-kube-api-access-rxpsj\") pod \"certified-operators-r9dc9\" (UID: \"9759ee01-8377-4d37-b262-dd591a841bd2\") " pod="openshift-marketplace/certified-operators-r9dc9" Dec 03 16:49:35 crc kubenswrapper[4998]: I1203 16:49:35.536133 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9759ee01-8377-4d37-b262-dd591a841bd2-utilities\") pod \"certified-operators-r9dc9\" (UID: \"9759ee01-8377-4d37-b262-dd591a841bd2\") " pod="openshift-marketplace/certified-operators-r9dc9" Dec 03 16:49:35 crc kubenswrapper[4998]: I1203 16:49:35.536202 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9759ee01-8377-4d37-b262-dd591a841bd2-catalog-content\") pod \"certified-operators-r9dc9\" (UID: \"9759ee01-8377-4d37-b262-dd591a841bd2\") " pod="openshift-marketplace/certified-operators-r9dc9" Dec 03 16:49:35 crc kubenswrapper[4998]: I1203 16:49:35.536275 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxpsj\" (UniqueName: \"kubernetes.io/projected/9759ee01-8377-4d37-b262-dd591a841bd2-kube-api-access-rxpsj\") pod \"certified-operators-r9dc9\" (UID: \"9759ee01-8377-4d37-b262-dd591a841bd2\") " pod="openshift-marketplace/certified-operators-r9dc9" Dec 03 16:49:35 crc kubenswrapper[4998]: I1203 16:49:35.536683 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9759ee01-8377-4d37-b262-dd591a841bd2-utilities\") pod \"certified-operators-r9dc9\" (UID: \"9759ee01-8377-4d37-b262-dd591a841bd2\") " pod="openshift-marketplace/certified-operators-r9dc9" Dec 03 16:49:35 crc kubenswrapper[4998]: I1203 16:49:35.536715 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9759ee01-8377-4d37-b262-dd591a841bd2-catalog-content\") pod \"certified-operators-r9dc9\" (UID: \"9759ee01-8377-4d37-b262-dd591a841bd2\") " pod="openshift-marketplace/certified-operators-r9dc9" Dec 03 16:49:35 crc kubenswrapper[4998]: I1203 16:49:35.556424 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxpsj\" (UniqueName: \"kubernetes.io/projected/9759ee01-8377-4d37-b262-dd591a841bd2-kube-api-access-rxpsj\") pod \"certified-operators-r9dc9\" (UID: \"9759ee01-8377-4d37-b262-dd591a841bd2\") " pod="openshift-marketplace/certified-operators-r9dc9" Dec 03 16:49:35 crc kubenswrapper[4998]: I1203 16:49:35.706647 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r9dc9" Dec 03 16:49:36 crc kubenswrapper[4998]: I1203 16:49:36.251675 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r9dc9"] Dec 03 16:49:36 crc kubenswrapper[4998]: I1203 16:49:36.659028 4998 generic.go:334] "Generic (PLEG): container finished" podID="9759ee01-8377-4d37-b262-dd591a841bd2" containerID="7638ba00fabea61be6f12d5ec583b7032c9fc9617907a9fe22f82fb4cfcedda1" exitCode=0 Dec 03 16:49:36 crc kubenswrapper[4998]: I1203 16:49:36.659085 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r9dc9" event={"ID":"9759ee01-8377-4d37-b262-dd591a841bd2","Type":"ContainerDied","Data":"7638ba00fabea61be6f12d5ec583b7032c9fc9617907a9fe22f82fb4cfcedda1"} Dec 03 16:49:36 crc kubenswrapper[4998]: I1203 16:49:36.659808 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r9dc9" event={"ID":"9759ee01-8377-4d37-b262-dd591a841bd2","Type":"ContainerStarted","Data":"37480be975f022254ac523128f97fa6233af723169305ebc2373e8d62142c2c0"} Dec 03 16:49:37 crc kubenswrapper[4998]: I1203 16:49:37.670427 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r9dc9" event={"ID":"9759ee01-8377-4d37-b262-dd591a841bd2","Type":"ContainerStarted","Data":"6b53de64c4335edfc9bc3a0070389798940ce3799322db1f8fb4a0ceb6272db3"} Dec 03 16:49:38 crc kubenswrapper[4998]: I1203 16:49:38.680625 4998 generic.go:334] "Generic (PLEG): container finished" podID="9759ee01-8377-4d37-b262-dd591a841bd2" containerID="6b53de64c4335edfc9bc3a0070389798940ce3799322db1f8fb4a0ceb6272db3" exitCode=0 Dec 03 16:49:38 crc kubenswrapper[4998]: I1203 16:49:38.680698 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r9dc9" event={"ID":"9759ee01-8377-4d37-b262-dd591a841bd2","Type":"ContainerDied","Data":"6b53de64c4335edfc9bc3a0070389798940ce3799322db1f8fb4a0ceb6272db3"} Dec 03 16:49:40 crc kubenswrapper[4998]: I1203 16:49:40.703483 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r9dc9" event={"ID":"9759ee01-8377-4d37-b262-dd591a841bd2","Type":"ContainerStarted","Data":"f7cf1d76d0003dbe4d23f73ddefcebb0ed42fb164be9b27a976dde064a8d2eda"} Dec 03 16:49:40 crc kubenswrapper[4998]: I1203 16:49:40.732279 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r9dc9" podStartSLOduration=2.588945334 podStartE2EDuration="5.732261206s" podCreationTimestamp="2025-12-03 16:49:35 +0000 UTC" firstStartedPulling="2025-12-03 16:49:36.660813927 +0000 UTC m=+2755.272514150" lastFinishedPulling="2025-12-03 16:49:39.804129799 +0000 UTC m=+2758.415830022" observedRunningTime="2025-12-03 16:49:40.721589744 +0000 UTC m=+2759.333289967" watchObservedRunningTime="2025-12-03 16:49:40.732261206 +0000 UTC m=+2759.343961429" Dec 03 16:49:45 crc kubenswrapper[4998]: I1203 16:49:45.707019 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r9dc9" Dec 03 16:49:45 crc kubenswrapper[4998]: I1203 16:49:45.707603 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r9dc9" Dec 03 16:49:45 crc kubenswrapper[4998]: I1203 16:49:45.758037 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r9dc9" Dec 03 16:49:45 crc kubenswrapper[4998]: I1203 16:49:45.809030 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r9dc9" Dec 03 16:49:45 crc kubenswrapper[4998]: I1203 16:49:45.996035 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r9dc9"] Dec 03 16:49:47 crc kubenswrapper[4998]: I1203 16:49:47.767570 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r9dc9" podUID="9759ee01-8377-4d37-b262-dd591a841bd2" containerName="registry-server" containerID="cri-o://f7cf1d76d0003dbe4d23f73ddefcebb0ed42fb164be9b27a976dde064a8d2eda" gracePeriod=2 Dec 03 16:49:48 crc kubenswrapper[4998]: I1203 16:49:48.783998 4998 generic.go:334] "Generic (PLEG): container finished" podID="9759ee01-8377-4d37-b262-dd591a841bd2" containerID="f7cf1d76d0003dbe4d23f73ddefcebb0ed42fb164be9b27a976dde064a8d2eda" exitCode=0 Dec 03 16:49:48 crc kubenswrapper[4998]: I1203 16:49:48.784097 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r9dc9" event={"ID":"9759ee01-8377-4d37-b262-dd591a841bd2","Type":"ContainerDied","Data":"f7cf1d76d0003dbe4d23f73ddefcebb0ed42fb164be9b27a976dde064a8d2eda"} Dec 03 16:49:49 crc kubenswrapper[4998]: I1203 16:49:49.435900 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r9dc9" Dec 03 16:49:49 crc kubenswrapper[4998]: I1203 16:49:49.514200 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9759ee01-8377-4d37-b262-dd591a841bd2-utilities\") pod \"9759ee01-8377-4d37-b262-dd591a841bd2\" (UID: \"9759ee01-8377-4d37-b262-dd591a841bd2\") " Dec 03 16:49:49 crc kubenswrapper[4998]: I1203 16:49:49.514274 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxpsj\" (UniqueName: \"kubernetes.io/projected/9759ee01-8377-4d37-b262-dd591a841bd2-kube-api-access-rxpsj\") pod \"9759ee01-8377-4d37-b262-dd591a841bd2\" (UID: \"9759ee01-8377-4d37-b262-dd591a841bd2\") " Dec 03 16:49:49 crc kubenswrapper[4998]: I1203 16:49:49.514405 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9759ee01-8377-4d37-b262-dd591a841bd2-catalog-content\") pod \"9759ee01-8377-4d37-b262-dd591a841bd2\" (UID: \"9759ee01-8377-4d37-b262-dd591a841bd2\") " Dec 03 16:49:49 crc kubenswrapper[4998]: I1203 16:49:49.515432 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9759ee01-8377-4d37-b262-dd591a841bd2-utilities" (OuterVolumeSpecName: "utilities") pod "9759ee01-8377-4d37-b262-dd591a841bd2" (UID: "9759ee01-8377-4d37-b262-dd591a841bd2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:49:49 crc kubenswrapper[4998]: I1203 16:49:49.521939 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9759ee01-8377-4d37-b262-dd591a841bd2-kube-api-access-rxpsj" (OuterVolumeSpecName: "kube-api-access-rxpsj") pod "9759ee01-8377-4d37-b262-dd591a841bd2" (UID: "9759ee01-8377-4d37-b262-dd591a841bd2"). InnerVolumeSpecName "kube-api-access-rxpsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:49:49 crc kubenswrapper[4998]: I1203 16:49:49.565875 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9759ee01-8377-4d37-b262-dd591a841bd2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9759ee01-8377-4d37-b262-dd591a841bd2" (UID: "9759ee01-8377-4d37-b262-dd591a841bd2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:49:49 crc kubenswrapper[4998]: I1203 16:49:49.617188 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9759ee01-8377-4d37-b262-dd591a841bd2-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:49:49 crc kubenswrapper[4998]: I1203 16:49:49.617232 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxpsj\" (UniqueName: \"kubernetes.io/projected/9759ee01-8377-4d37-b262-dd591a841bd2-kube-api-access-rxpsj\") on node \"crc\" DevicePath \"\"" Dec 03 16:49:49 crc kubenswrapper[4998]: I1203 16:49:49.617247 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9759ee01-8377-4d37-b262-dd591a841bd2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:49:49 crc kubenswrapper[4998]: I1203 16:49:49.798665 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r9dc9" event={"ID":"9759ee01-8377-4d37-b262-dd591a841bd2","Type":"ContainerDied","Data":"37480be975f022254ac523128f97fa6233af723169305ebc2373e8d62142c2c0"} Dec 03 16:49:49 crc kubenswrapper[4998]: I1203 16:49:49.798741 4998 scope.go:117] "RemoveContainer" containerID="f7cf1d76d0003dbe4d23f73ddefcebb0ed42fb164be9b27a976dde064a8d2eda" Dec 03 16:49:49 crc kubenswrapper[4998]: I1203 16:49:49.798769 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r9dc9" Dec 03 16:49:49 crc kubenswrapper[4998]: I1203 16:49:49.821505 4998 scope.go:117] "RemoveContainer" containerID="6b53de64c4335edfc9bc3a0070389798940ce3799322db1f8fb4a0ceb6272db3" Dec 03 16:49:49 crc kubenswrapper[4998]: I1203 16:49:49.832274 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r9dc9"] Dec 03 16:49:49 crc kubenswrapper[4998]: I1203 16:49:49.849812 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r9dc9"] Dec 03 16:49:49 crc kubenswrapper[4998]: I1203 16:49:49.854591 4998 scope.go:117] "RemoveContainer" containerID="7638ba00fabea61be6f12d5ec583b7032c9fc9617907a9fe22f82fb4cfcedda1" Dec 03 16:49:51 crc kubenswrapper[4998]: I1203 16:49:51.704709 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9759ee01-8377-4d37-b262-dd591a841bd2" path="/var/lib/kubelet/pods/9759ee01-8377-4d37-b262-dd591a841bd2/volumes" Dec 03 16:50:06 crc kubenswrapper[4998]: I1203 16:50:06.730069 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4bbxw"] Dec 03 16:50:06 crc kubenswrapper[4998]: E1203 16:50:06.730946 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9759ee01-8377-4d37-b262-dd591a841bd2" containerName="extract-content" Dec 03 16:50:06 crc kubenswrapper[4998]: I1203 16:50:06.730965 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="9759ee01-8377-4d37-b262-dd591a841bd2" containerName="extract-content" Dec 03 16:50:06 crc kubenswrapper[4998]: E1203 16:50:06.730984 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9759ee01-8377-4d37-b262-dd591a841bd2" containerName="registry-server" Dec 03 16:50:06 crc kubenswrapper[4998]: I1203 16:50:06.730991 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="9759ee01-8377-4d37-b262-dd591a841bd2" containerName="registry-server" Dec 03 16:50:06 crc kubenswrapper[4998]: E1203 16:50:06.731020 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9759ee01-8377-4d37-b262-dd591a841bd2" containerName="extract-utilities" Dec 03 16:50:06 crc kubenswrapper[4998]: I1203 16:50:06.731028 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="9759ee01-8377-4d37-b262-dd591a841bd2" containerName="extract-utilities" Dec 03 16:50:06 crc kubenswrapper[4998]: I1203 16:50:06.731282 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="9759ee01-8377-4d37-b262-dd591a841bd2" containerName="registry-server" Dec 03 16:50:06 crc kubenswrapper[4998]: I1203 16:50:06.733156 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4bbxw" Dec 03 16:50:06 crc kubenswrapper[4998]: I1203 16:50:06.755040 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4bbxw"] Dec 03 16:50:06 crc kubenswrapper[4998]: I1203 16:50:06.766290 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcb6db43-2b68-4707-8211-08bbc091789f-utilities\") pod \"community-operators-4bbxw\" (UID: \"dcb6db43-2b68-4707-8211-08bbc091789f\") " pod="openshift-marketplace/community-operators-4bbxw" Dec 03 16:50:06 crc kubenswrapper[4998]: I1203 16:50:06.766347 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcb6db43-2b68-4707-8211-08bbc091789f-catalog-content\") pod \"community-operators-4bbxw\" (UID: \"dcb6db43-2b68-4707-8211-08bbc091789f\") " pod="openshift-marketplace/community-operators-4bbxw" Dec 03 16:50:06 crc kubenswrapper[4998]: I1203 16:50:06.766525 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pswn2\" (UniqueName: \"kubernetes.io/projected/dcb6db43-2b68-4707-8211-08bbc091789f-kube-api-access-pswn2\") pod \"community-operators-4bbxw\" (UID: \"dcb6db43-2b68-4707-8211-08bbc091789f\") " pod="openshift-marketplace/community-operators-4bbxw" Dec 03 16:50:06 crc kubenswrapper[4998]: I1203 16:50:06.868723 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pswn2\" (UniqueName: \"kubernetes.io/projected/dcb6db43-2b68-4707-8211-08bbc091789f-kube-api-access-pswn2\") pod \"community-operators-4bbxw\" (UID: \"dcb6db43-2b68-4707-8211-08bbc091789f\") " pod="openshift-marketplace/community-operators-4bbxw" Dec 03 16:50:06 crc kubenswrapper[4998]: I1203 16:50:06.868886 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcb6db43-2b68-4707-8211-08bbc091789f-utilities\") pod \"community-operators-4bbxw\" (UID: \"dcb6db43-2b68-4707-8211-08bbc091789f\") " pod="openshift-marketplace/community-operators-4bbxw" Dec 03 16:50:06 crc kubenswrapper[4998]: I1203 16:50:06.868908 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcb6db43-2b68-4707-8211-08bbc091789f-catalog-content\") pod \"community-operators-4bbxw\" (UID: \"dcb6db43-2b68-4707-8211-08bbc091789f\") " pod="openshift-marketplace/community-operators-4bbxw" Dec 03 16:50:06 crc kubenswrapper[4998]: I1203 16:50:06.869298 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcb6db43-2b68-4707-8211-08bbc091789f-utilities\") pod \"community-operators-4bbxw\" (UID: \"dcb6db43-2b68-4707-8211-08bbc091789f\") " pod="openshift-marketplace/community-operators-4bbxw" Dec 03 16:50:06 crc kubenswrapper[4998]: I1203 16:50:06.869356 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcb6db43-2b68-4707-8211-08bbc091789f-catalog-content\") pod \"community-operators-4bbxw\" (UID: \"dcb6db43-2b68-4707-8211-08bbc091789f\") " pod="openshift-marketplace/community-operators-4bbxw" Dec 03 16:50:06 crc kubenswrapper[4998]: I1203 16:50:06.894653 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pswn2\" (UniqueName: \"kubernetes.io/projected/dcb6db43-2b68-4707-8211-08bbc091789f-kube-api-access-pswn2\") pod \"community-operators-4bbxw\" (UID: \"dcb6db43-2b68-4707-8211-08bbc091789f\") " pod="openshift-marketplace/community-operators-4bbxw" Dec 03 16:50:07 crc kubenswrapper[4998]: I1203 16:50:07.059586 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4bbxw" Dec 03 16:50:07 crc kubenswrapper[4998]: I1203 16:50:07.612546 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4bbxw"] Dec 03 16:50:07 crc kubenswrapper[4998]: I1203 16:50:07.967202 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4bbxw" event={"ID":"dcb6db43-2b68-4707-8211-08bbc091789f","Type":"ContainerStarted","Data":"16dd6e1712b4de8710efc7b0bce51a89a8cf15502b3d0dbccc0b8d30004a6f85"} Dec 03 16:50:09 crc kubenswrapper[4998]: I1203 16:50:09.985074 4998 generic.go:334] "Generic (PLEG): container finished" podID="dcb6db43-2b68-4707-8211-08bbc091789f" containerID="ce243aedb2832a8ef0fa231a068618432fafb0cf5c721209c66371699571b203" exitCode=0 Dec 03 16:50:09 crc kubenswrapper[4998]: I1203 16:50:09.985168 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4bbxw" event={"ID":"dcb6db43-2b68-4707-8211-08bbc091789f","Type":"ContainerDied","Data":"ce243aedb2832a8ef0fa231a068618432fafb0cf5c721209c66371699571b203"} Dec 03 16:50:10 crc kubenswrapper[4998]: I1203 16:50:10.996304 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4bbxw" event={"ID":"dcb6db43-2b68-4707-8211-08bbc091789f","Type":"ContainerStarted","Data":"4e1e96470f972d0079408ed2f93290164cc09d831e6d9c4b1e07bddd0a71b40f"} Dec 03 16:50:12 crc kubenswrapper[4998]: I1203 16:50:12.008298 4998 generic.go:334] "Generic (PLEG): container finished" podID="dcb6db43-2b68-4707-8211-08bbc091789f" containerID="4e1e96470f972d0079408ed2f93290164cc09d831e6d9c4b1e07bddd0a71b40f" exitCode=0 Dec 03 16:50:12 crc kubenswrapper[4998]: I1203 16:50:12.008355 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4bbxw" event={"ID":"dcb6db43-2b68-4707-8211-08bbc091789f","Type":"ContainerDied","Data":"4e1e96470f972d0079408ed2f93290164cc09d831e6d9c4b1e07bddd0a71b40f"} Dec 03 16:50:13 crc kubenswrapper[4998]: I1203 16:50:13.020028 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4bbxw" event={"ID":"dcb6db43-2b68-4707-8211-08bbc091789f","Type":"ContainerStarted","Data":"1c568faf116da9e487f562bc82e61d92c932639f883a70479913b052d7879743"} Dec 03 16:50:13 crc kubenswrapper[4998]: I1203 16:50:13.048630 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4bbxw" podStartSLOduration=4.637233003 podStartE2EDuration="7.048609109s" podCreationTimestamp="2025-12-03 16:50:06 +0000 UTC" firstStartedPulling="2025-12-03 16:50:09.98700745 +0000 UTC m=+2788.598707683" lastFinishedPulling="2025-12-03 16:50:12.398383526 +0000 UTC m=+2791.010083789" observedRunningTime="2025-12-03 16:50:13.037986118 +0000 UTC m=+2791.649686351" watchObservedRunningTime="2025-12-03 16:50:13.048609109 +0000 UTC m=+2791.660309342" Dec 03 16:50:17 crc kubenswrapper[4998]: I1203 16:50:17.060535 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4bbxw" Dec 03 16:50:17 crc kubenswrapper[4998]: I1203 16:50:17.061077 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4bbxw" Dec 03 16:50:17 crc kubenswrapper[4998]: I1203 16:50:17.107791 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4bbxw" Dec 03 16:50:18 crc kubenswrapper[4998]: I1203 16:50:18.115883 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4bbxw" Dec 03 16:50:18 crc kubenswrapper[4998]: I1203 16:50:18.176378 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4bbxw"] Dec 03 16:50:20 crc kubenswrapper[4998]: I1203 16:50:20.084038 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4bbxw" podUID="dcb6db43-2b68-4707-8211-08bbc091789f" containerName="registry-server" containerID="cri-o://1c568faf116da9e487f562bc82e61d92c932639f883a70479913b052d7879743" gracePeriod=2 Dec 03 16:50:20 crc kubenswrapper[4998]: I1203 16:50:20.535198 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4bbxw" Dec 03 16:50:20 crc kubenswrapper[4998]: I1203 16:50:20.567307 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcb6db43-2b68-4707-8211-08bbc091789f-utilities\") pod \"dcb6db43-2b68-4707-8211-08bbc091789f\" (UID: \"dcb6db43-2b68-4707-8211-08bbc091789f\") " Dec 03 16:50:20 crc kubenswrapper[4998]: I1203 16:50:20.567408 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pswn2\" (UniqueName: \"kubernetes.io/projected/dcb6db43-2b68-4707-8211-08bbc091789f-kube-api-access-pswn2\") pod \"dcb6db43-2b68-4707-8211-08bbc091789f\" (UID: \"dcb6db43-2b68-4707-8211-08bbc091789f\") " Dec 03 16:50:20 crc kubenswrapper[4998]: I1203 16:50:20.567447 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcb6db43-2b68-4707-8211-08bbc091789f-catalog-content\") pod \"dcb6db43-2b68-4707-8211-08bbc091789f\" (UID: \"dcb6db43-2b68-4707-8211-08bbc091789f\") " Dec 03 16:50:20 crc kubenswrapper[4998]: I1203 16:50:20.569536 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcb6db43-2b68-4707-8211-08bbc091789f-utilities" (OuterVolumeSpecName: "utilities") pod "dcb6db43-2b68-4707-8211-08bbc091789f" (UID: "dcb6db43-2b68-4707-8211-08bbc091789f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:50:20 crc kubenswrapper[4998]: I1203 16:50:20.575576 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcb6db43-2b68-4707-8211-08bbc091789f-kube-api-access-pswn2" (OuterVolumeSpecName: "kube-api-access-pswn2") pod "dcb6db43-2b68-4707-8211-08bbc091789f" (UID: "dcb6db43-2b68-4707-8211-08bbc091789f"). InnerVolumeSpecName "kube-api-access-pswn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:50:20 crc kubenswrapper[4998]: I1203 16:50:20.618231 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcb6db43-2b68-4707-8211-08bbc091789f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dcb6db43-2b68-4707-8211-08bbc091789f" (UID: "dcb6db43-2b68-4707-8211-08bbc091789f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:50:20 crc kubenswrapper[4998]: I1203 16:50:20.671101 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcb6db43-2b68-4707-8211-08bbc091789f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:50:20 crc kubenswrapper[4998]: I1203 16:50:20.671138 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pswn2\" (UniqueName: \"kubernetes.io/projected/dcb6db43-2b68-4707-8211-08bbc091789f-kube-api-access-pswn2\") on node \"crc\" DevicePath \"\"" Dec 03 16:50:20 crc kubenswrapper[4998]: I1203 16:50:20.671151 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcb6db43-2b68-4707-8211-08bbc091789f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:50:21 crc kubenswrapper[4998]: I1203 16:50:21.103133 4998 generic.go:334] "Generic (PLEG): container finished" podID="dcb6db43-2b68-4707-8211-08bbc091789f" containerID="1c568faf116da9e487f562bc82e61d92c932639f883a70479913b052d7879743" exitCode=0 Dec 03 16:50:21 crc kubenswrapper[4998]: I1203 16:50:21.103488 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4bbxw" event={"ID":"dcb6db43-2b68-4707-8211-08bbc091789f","Type":"ContainerDied","Data":"1c568faf116da9e487f562bc82e61d92c932639f883a70479913b052d7879743"} Dec 03 16:50:21 crc kubenswrapper[4998]: I1203 16:50:21.103521 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4bbxw" event={"ID":"dcb6db43-2b68-4707-8211-08bbc091789f","Type":"ContainerDied","Data":"16dd6e1712b4de8710efc7b0bce51a89a8cf15502b3d0dbccc0b8d30004a6f85"} Dec 03 16:50:21 crc kubenswrapper[4998]: I1203 16:50:21.103542 4998 scope.go:117] "RemoveContainer" containerID="1c568faf116da9e487f562bc82e61d92c932639f883a70479913b052d7879743" Dec 03 16:50:21 crc kubenswrapper[4998]: I1203 16:50:21.103695 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4bbxw" Dec 03 16:50:21 crc kubenswrapper[4998]: I1203 16:50:21.139688 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4bbxw"] Dec 03 16:50:21 crc kubenswrapper[4998]: I1203 16:50:21.146154 4998 scope.go:117] "RemoveContainer" containerID="4e1e96470f972d0079408ed2f93290164cc09d831e6d9c4b1e07bddd0a71b40f" Dec 03 16:50:21 crc kubenswrapper[4998]: I1203 16:50:21.148972 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4bbxw"] Dec 03 16:50:21 crc kubenswrapper[4998]: I1203 16:50:21.176607 4998 scope.go:117] "RemoveContainer" containerID="ce243aedb2832a8ef0fa231a068618432fafb0cf5c721209c66371699571b203" Dec 03 16:50:21 crc kubenswrapper[4998]: I1203 16:50:21.226322 4998 scope.go:117] "RemoveContainer" containerID="1c568faf116da9e487f562bc82e61d92c932639f883a70479913b052d7879743" Dec 03 16:50:21 crc kubenswrapper[4998]: E1203 16:50:21.226962 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c568faf116da9e487f562bc82e61d92c932639f883a70479913b052d7879743\": container with ID starting with 1c568faf116da9e487f562bc82e61d92c932639f883a70479913b052d7879743 not found: ID does not exist" containerID="1c568faf116da9e487f562bc82e61d92c932639f883a70479913b052d7879743" Dec 03 16:50:21 crc kubenswrapper[4998]: I1203 16:50:21.226999 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c568faf116da9e487f562bc82e61d92c932639f883a70479913b052d7879743"} err="failed to get container status \"1c568faf116da9e487f562bc82e61d92c932639f883a70479913b052d7879743\": rpc error: code = NotFound desc = could not find container \"1c568faf116da9e487f562bc82e61d92c932639f883a70479913b052d7879743\": container with ID starting with 1c568faf116da9e487f562bc82e61d92c932639f883a70479913b052d7879743 not found: ID does not exist" Dec 03 16:50:21 crc kubenswrapper[4998]: I1203 16:50:21.227025 4998 scope.go:117] "RemoveContainer" containerID="4e1e96470f972d0079408ed2f93290164cc09d831e6d9c4b1e07bddd0a71b40f" Dec 03 16:50:21 crc kubenswrapper[4998]: E1203 16:50:21.227265 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e1e96470f972d0079408ed2f93290164cc09d831e6d9c4b1e07bddd0a71b40f\": container with ID starting with 4e1e96470f972d0079408ed2f93290164cc09d831e6d9c4b1e07bddd0a71b40f not found: ID does not exist" containerID="4e1e96470f972d0079408ed2f93290164cc09d831e6d9c4b1e07bddd0a71b40f" Dec 03 16:50:21 crc kubenswrapper[4998]: I1203 16:50:21.227300 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e1e96470f972d0079408ed2f93290164cc09d831e6d9c4b1e07bddd0a71b40f"} err="failed to get container status \"4e1e96470f972d0079408ed2f93290164cc09d831e6d9c4b1e07bddd0a71b40f\": rpc error: code = NotFound desc = could not find container \"4e1e96470f972d0079408ed2f93290164cc09d831e6d9c4b1e07bddd0a71b40f\": container with ID starting with 4e1e96470f972d0079408ed2f93290164cc09d831e6d9c4b1e07bddd0a71b40f not found: ID does not exist" Dec 03 16:50:21 crc kubenswrapper[4998]: I1203 16:50:21.227318 4998 scope.go:117] "RemoveContainer" containerID="ce243aedb2832a8ef0fa231a068618432fafb0cf5c721209c66371699571b203" Dec 03 16:50:21 crc kubenswrapper[4998]: E1203 16:50:21.227533 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce243aedb2832a8ef0fa231a068618432fafb0cf5c721209c66371699571b203\": container with ID starting with ce243aedb2832a8ef0fa231a068618432fafb0cf5c721209c66371699571b203 not found: ID does not exist" containerID="ce243aedb2832a8ef0fa231a068618432fafb0cf5c721209c66371699571b203" Dec 03 16:50:21 crc kubenswrapper[4998]: I1203 16:50:21.227557 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce243aedb2832a8ef0fa231a068618432fafb0cf5c721209c66371699571b203"} err="failed to get container status \"ce243aedb2832a8ef0fa231a068618432fafb0cf5c721209c66371699571b203\": rpc error: code = NotFound desc = could not find container \"ce243aedb2832a8ef0fa231a068618432fafb0cf5c721209c66371699571b203\": container with ID starting with ce243aedb2832a8ef0fa231a068618432fafb0cf5c721209c66371699571b203 not found: ID does not exist" Dec 03 16:50:21 crc kubenswrapper[4998]: I1203 16:50:21.688329 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcb6db43-2b68-4707-8211-08bbc091789f" path="/var/lib/kubelet/pods/dcb6db43-2b68-4707-8211-08bbc091789f/volumes" Dec 03 16:50:27 crc kubenswrapper[4998]: I1203 16:50:27.111237 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:50:27 crc kubenswrapper[4998]: I1203 16:50:27.112174 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:50:57 crc kubenswrapper[4998]: I1203 16:50:57.112094 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:50:57 crc kubenswrapper[4998]: I1203 16:50:57.113023 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:51:27 crc kubenswrapper[4998]: I1203 16:51:27.111340 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:51:27 crc kubenswrapper[4998]: I1203 16:51:27.111887 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:51:27 crc kubenswrapper[4998]: I1203 16:51:27.111933 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:51:27 crc kubenswrapper[4998]: I1203 16:51:27.112991 4998 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096"} pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 16:51:27 crc kubenswrapper[4998]: I1203 16:51:27.113157 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" containerID="cri-o://21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" gracePeriod=600 Dec 03 16:51:27 crc kubenswrapper[4998]: E1203 16:51:27.262535 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:51:27 crc kubenswrapper[4998]: I1203 16:51:27.782269 4998 generic.go:334] "Generic (PLEG): container finished" podID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" exitCode=0 Dec 03 16:51:27 crc kubenswrapper[4998]: I1203 16:51:27.782388 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerDied","Data":"21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096"} Dec 03 16:51:27 crc kubenswrapper[4998]: I1203 16:51:27.782499 4998 scope.go:117] "RemoveContainer" containerID="65e608d68983552183ea47364a0c23e120fed4872f3ce284dfdd07db6b645cc3" Dec 03 16:51:27 crc kubenswrapper[4998]: I1203 16:51:27.784394 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:51:27 crc kubenswrapper[4998]: E1203 16:51:27.785588 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:51:28 crc kubenswrapper[4998]: I1203 16:51:28.794606 4998 generic.go:334] "Generic (PLEG): container finished" podID="5af60798-c3fe-4b82-9732-4cfd4c442549" containerID="d35a6171cb2a45a70783e1f5f6031a2297c93bc0b6e3be537740b47b3b0daa83" exitCode=0 Dec 03 16:51:28 crc kubenswrapper[4998]: I1203 16:51:28.794704 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" event={"ID":"5af60798-c3fe-4b82-9732-4cfd4c442549","Type":"ContainerDied","Data":"d35a6171cb2a45a70783e1f5f6031a2297c93bc0b6e3be537740b47b3b0daa83"} Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.217031 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.329370 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-combined-ca-bundle\") pod \"5af60798-c3fe-4b82-9732-4cfd4c442549\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.329412 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-migration-ssh-key-0\") pod \"5af60798-c3fe-4b82-9732-4cfd4c442549\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.329469 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-cell1-compute-config-0\") pod \"5af60798-c3fe-4b82-9732-4cfd4c442549\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.329499 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-migration-ssh-key-1\") pod \"5af60798-c3fe-4b82-9732-4cfd4c442549\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.329546 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-inventory\") pod \"5af60798-c3fe-4b82-9732-4cfd4c442549\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.329614 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-ssh-key\") pod \"5af60798-c3fe-4b82-9732-4cfd4c442549\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.329636 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-extra-config-0\") pod \"5af60798-c3fe-4b82-9732-4cfd4c442549\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.330266 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-cell1-compute-config-1\") pod \"5af60798-c3fe-4b82-9732-4cfd4c442549\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.330316 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4hfp\" (UniqueName: \"kubernetes.io/projected/5af60798-c3fe-4b82-9732-4cfd4c442549-kube-api-access-n4hfp\") pod \"5af60798-c3fe-4b82-9732-4cfd4c442549\" (UID: \"5af60798-c3fe-4b82-9732-4cfd4c442549\") " Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.336117 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5af60798-c3fe-4b82-9732-4cfd4c442549-kube-api-access-n4hfp" (OuterVolumeSpecName: "kube-api-access-n4hfp") pod "5af60798-c3fe-4b82-9732-4cfd4c442549" (UID: "5af60798-c3fe-4b82-9732-4cfd4c442549"). InnerVolumeSpecName "kube-api-access-n4hfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.341823 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "5af60798-c3fe-4b82-9732-4cfd4c442549" (UID: "5af60798-c3fe-4b82-9732-4cfd4c442549"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.360960 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "5af60798-c3fe-4b82-9732-4cfd4c442549" (UID: "5af60798-c3fe-4b82-9732-4cfd4c442549"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.362365 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "5af60798-c3fe-4b82-9732-4cfd4c442549" (UID: "5af60798-c3fe-4b82-9732-4cfd4c442549"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.364168 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "5af60798-c3fe-4b82-9732-4cfd4c442549" (UID: "5af60798-c3fe-4b82-9732-4cfd4c442549"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.364505 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-inventory" (OuterVolumeSpecName: "inventory") pod "5af60798-c3fe-4b82-9732-4cfd4c442549" (UID: "5af60798-c3fe-4b82-9732-4cfd4c442549"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.366078 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "5af60798-c3fe-4b82-9732-4cfd4c442549" (UID: "5af60798-c3fe-4b82-9732-4cfd4c442549"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.366924 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "5af60798-c3fe-4b82-9732-4cfd4c442549" (UID: "5af60798-c3fe-4b82-9732-4cfd4c442549"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.377379 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5af60798-c3fe-4b82-9732-4cfd4c442549" (UID: "5af60798-c3fe-4b82-9732-4cfd4c442549"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.432886 4998 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.433147 4998 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.433215 4998 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.433289 4998 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.433342 4998 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.433403 4998 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.433464 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4hfp\" (UniqueName: \"kubernetes.io/projected/5af60798-c3fe-4b82-9732-4cfd4c442549-kube-api-access-n4hfp\") on node \"crc\" DevicePath \"\"" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.433529 4998 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.433581 4998 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/5af60798-c3fe-4b82-9732-4cfd4c442549-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.816255 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" event={"ID":"5af60798-c3fe-4b82-9732-4cfd4c442549","Type":"ContainerDied","Data":"a49cdd635df653fab64a83df37257dd4cb85c54ecc5357bd6dcc91b73265238b"} Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.816669 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a49cdd635df653fab64a83df37257dd4cb85c54ecc5357bd6dcc91b73265238b" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.816468 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-w6l7l" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.926060 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2"] Dec 03 16:51:30 crc kubenswrapper[4998]: E1203 16:51:30.926516 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcb6db43-2b68-4707-8211-08bbc091789f" containerName="registry-server" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.926533 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcb6db43-2b68-4707-8211-08bbc091789f" containerName="registry-server" Dec 03 16:51:30 crc kubenswrapper[4998]: E1203 16:51:30.926546 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af60798-c3fe-4b82-9732-4cfd4c442549" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.926554 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af60798-c3fe-4b82-9732-4cfd4c442549" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 03 16:51:30 crc kubenswrapper[4998]: E1203 16:51:30.926580 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcb6db43-2b68-4707-8211-08bbc091789f" containerName="extract-content" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.926587 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcb6db43-2b68-4707-8211-08bbc091789f" containerName="extract-content" Dec 03 16:51:30 crc kubenswrapper[4998]: E1203 16:51:30.926603 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcb6db43-2b68-4707-8211-08bbc091789f" containerName="extract-utilities" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.926610 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcb6db43-2b68-4707-8211-08bbc091789f" containerName="extract-utilities" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.926804 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="5af60798-c3fe-4b82-9732-4cfd4c442549" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.926824 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcb6db43-2b68-4707-8211-08bbc091789f" containerName="registry-server" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.927477 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.933029 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.933145 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.933029 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.933343 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kcj9x" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.933573 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 16:51:30 crc kubenswrapper[4998]: I1203 16:51:30.960295 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2"] Dec 03 16:51:31 crc kubenswrapper[4998]: I1203 16:51:31.049388 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" Dec 03 16:51:31 crc kubenswrapper[4998]: I1203 16:51:31.049503 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" Dec 03 16:51:31 crc kubenswrapper[4998]: I1203 16:51:31.049554 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" Dec 03 16:51:31 crc kubenswrapper[4998]: I1203 16:51:31.049576 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" Dec 03 16:51:31 crc kubenswrapper[4998]: I1203 16:51:31.049600 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b28wt\" (UniqueName: \"kubernetes.io/projected/10d74042-5cd9-4733-9612-2357a2577f0a-kube-api-access-b28wt\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" Dec 03 16:51:31 crc kubenswrapper[4998]: I1203 16:51:31.049639 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" Dec 03 16:51:31 crc kubenswrapper[4998]: I1203 16:51:31.049657 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" Dec 03 16:51:31 crc kubenswrapper[4998]: I1203 16:51:31.151056 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" Dec 03 16:51:31 crc kubenswrapper[4998]: I1203 16:51:31.151125 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" Dec 03 16:51:31 crc kubenswrapper[4998]: I1203 16:51:31.151862 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" Dec 03 16:51:31 crc kubenswrapper[4998]: I1203 16:51:31.151915 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b28wt\" (UniqueName: \"kubernetes.io/projected/10d74042-5cd9-4733-9612-2357a2577f0a-kube-api-access-b28wt\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" Dec 03 16:51:31 crc kubenswrapper[4998]: I1203 16:51:31.151973 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" Dec 03 16:51:31 crc kubenswrapper[4998]: I1203 16:51:31.152002 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" Dec 03 16:51:31 crc kubenswrapper[4998]: I1203 16:51:31.152078 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" Dec 03 16:51:31 crc kubenswrapper[4998]: I1203 16:51:31.155282 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" Dec 03 16:51:31 crc kubenswrapper[4998]: I1203 16:51:31.155666 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" Dec 03 16:51:31 crc kubenswrapper[4998]: I1203 16:51:31.156058 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" Dec 03 16:51:31 crc kubenswrapper[4998]: I1203 16:51:31.156108 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" Dec 03 16:51:31 crc kubenswrapper[4998]: I1203 16:51:31.158427 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" Dec 03 16:51:31 crc kubenswrapper[4998]: I1203 16:51:31.161332 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" Dec 03 16:51:31 crc kubenswrapper[4998]: I1203 16:51:31.169293 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b28wt\" (UniqueName: \"kubernetes.io/projected/10d74042-5cd9-4733-9612-2357a2577f0a-kube-api-access-b28wt\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" Dec 03 16:51:31 crc kubenswrapper[4998]: I1203 16:51:31.251267 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" Dec 03 16:51:31 crc kubenswrapper[4998]: I1203 16:51:31.758241 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2"] Dec 03 16:51:31 crc kubenswrapper[4998]: I1203 16:51:31.825495 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" event={"ID":"10d74042-5cd9-4733-9612-2357a2577f0a","Type":"ContainerStarted","Data":"37fc94e07cb3730df8bcc3621d07fb6983bee732a94fa2da6ac191c5de59a43b"} Dec 03 16:51:32 crc kubenswrapper[4998]: I1203 16:51:32.837446 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" event={"ID":"10d74042-5cd9-4733-9612-2357a2577f0a","Type":"ContainerStarted","Data":"aee5fbb02e431dbfb6252731a809a25fc8573cdbc3557030c3dc5a73b0061e93"} Dec 03 16:51:32 crc kubenswrapper[4998]: I1203 16:51:32.859487 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" podStartSLOduration=2.384618686 podStartE2EDuration="2.859459548s" podCreationTimestamp="2025-12-03 16:51:30 +0000 UTC" firstStartedPulling="2025-12-03 16:51:31.767685229 +0000 UTC m=+2870.379385452" lastFinishedPulling="2025-12-03 16:51:32.242526081 +0000 UTC m=+2870.854226314" observedRunningTime="2025-12-03 16:51:32.858161536 +0000 UTC m=+2871.469861769" watchObservedRunningTime="2025-12-03 16:51:32.859459548 +0000 UTC m=+2871.471159811" Dec 03 16:51:39 crc kubenswrapper[4998]: I1203 16:51:39.678050 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:51:39 crc kubenswrapper[4998]: E1203 16:51:39.679132 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:51:52 crc kubenswrapper[4998]: I1203 16:51:52.678473 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:51:52 crc kubenswrapper[4998]: E1203 16:51:52.679639 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:52:04 crc kubenswrapper[4998]: I1203 16:52:04.678643 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:52:04 crc kubenswrapper[4998]: E1203 16:52:04.679351 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:52:18 crc kubenswrapper[4998]: I1203 16:52:18.677444 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:52:18 crc kubenswrapper[4998]: E1203 16:52:18.678299 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:52:30 crc kubenswrapper[4998]: I1203 16:52:30.677613 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:52:30 crc kubenswrapper[4998]: E1203 16:52:30.678683 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:52:43 crc kubenswrapper[4998]: I1203 16:52:43.677743 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:52:43 crc kubenswrapper[4998]: E1203 16:52:43.678896 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:52:55 crc kubenswrapper[4998]: I1203 16:52:55.677863 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:52:55 crc kubenswrapper[4998]: E1203 16:52:55.679688 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:53:09 crc kubenswrapper[4998]: I1203 16:53:09.677750 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:53:09 crc kubenswrapper[4998]: E1203 16:53:09.678639 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:53:21 crc kubenswrapper[4998]: I1203 16:53:21.677866 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:53:21 crc kubenswrapper[4998]: E1203 16:53:21.678706 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:53:32 crc kubenswrapper[4998]: I1203 16:53:32.678043 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:53:32 crc kubenswrapper[4998]: E1203 16:53:32.678974 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:53:46 crc kubenswrapper[4998]: I1203 16:53:46.678360 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:53:46 crc kubenswrapper[4998]: E1203 16:53:46.679129 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:53:49 crc kubenswrapper[4998]: I1203 16:53:49.314925 4998 generic.go:334] "Generic (PLEG): container finished" podID="10d74042-5cd9-4733-9612-2357a2577f0a" containerID="aee5fbb02e431dbfb6252731a809a25fc8573cdbc3557030c3dc5a73b0061e93" exitCode=0 Dec 03 16:53:49 crc kubenswrapper[4998]: I1203 16:53:49.314982 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" event={"ID":"10d74042-5cd9-4733-9612-2357a2577f0a","Type":"ContainerDied","Data":"aee5fbb02e431dbfb6252731a809a25fc8573cdbc3557030c3dc5a73b0061e93"} Dec 03 16:53:50 crc kubenswrapper[4998]: I1203 16:53:50.781732 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" Dec 03 16:53:50 crc kubenswrapper[4998]: I1203 16:53:50.848445 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-inventory\") pod \"10d74042-5cd9-4733-9612-2357a2577f0a\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " Dec 03 16:53:50 crc kubenswrapper[4998]: I1203 16:53:50.848506 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b28wt\" (UniqueName: \"kubernetes.io/projected/10d74042-5cd9-4733-9612-2357a2577f0a-kube-api-access-b28wt\") pod \"10d74042-5cd9-4733-9612-2357a2577f0a\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " Dec 03 16:53:50 crc kubenswrapper[4998]: I1203 16:53:50.848572 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-ceilometer-compute-config-data-0\") pod \"10d74042-5cd9-4733-9612-2357a2577f0a\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " Dec 03 16:53:50 crc kubenswrapper[4998]: I1203 16:53:50.848601 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-ssh-key\") pod \"10d74042-5cd9-4733-9612-2357a2577f0a\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " Dec 03 16:53:50 crc kubenswrapper[4998]: I1203 16:53:50.848662 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-ceilometer-compute-config-data-2\") pod \"10d74042-5cd9-4733-9612-2357a2577f0a\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " Dec 03 16:53:50 crc kubenswrapper[4998]: I1203 16:53:50.849205 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-telemetry-combined-ca-bundle\") pod \"10d74042-5cd9-4733-9612-2357a2577f0a\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " Dec 03 16:53:50 crc kubenswrapper[4998]: I1203 16:53:50.849295 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-ceilometer-compute-config-data-1\") pod \"10d74042-5cd9-4733-9612-2357a2577f0a\" (UID: \"10d74042-5cd9-4733-9612-2357a2577f0a\") " Dec 03 16:53:50 crc kubenswrapper[4998]: I1203 16:53:50.856132 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10d74042-5cd9-4733-9612-2357a2577f0a-kube-api-access-b28wt" (OuterVolumeSpecName: "kube-api-access-b28wt") pod "10d74042-5cd9-4733-9612-2357a2577f0a" (UID: "10d74042-5cd9-4733-9612-2357a2577f0a"). InnerVolumeSpecName "kube-api-access-b28wt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:53:50 crc kubenswrapper[4998]: I1203 16:53:50.868284 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "10d74042-5cd9-4733-9612-2357a2577f0a" (UID: "10d74042-5cd9-4733-9612-2357a2577f0a"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:53:50 crc kubenswrapper[4998]: I1203 16:53:50.880968 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "10d74042-5cd9-4733-9612-2357a2577f0a" (UID: "10d74042-5cd9-4733-9612-2357a2577f0a"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:53:50 crc kubenswrapper[4998]: I1203 16:53:50.886981 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-inventory" (OuterVolumeSpecName: "inventory") pod "10d74042-5cd9-4733-9612-2357a2577f0a" (UID: "10d74042-5cd9-4733-9612-2357a2577f0a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:53:50 crc kubenswrapper[4998]: I1203 16:53:50.887915 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "10d74042-5cd9-4733-9612-2357a2577f0a" (UID: "10d74042-5cd9-4733-9612-2357a2577f0a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:53:50 crc kubenswrapper[4998]: I1203 16:53:50.892566 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "10d74042-5cd9-4733-9612-2357a2577f0a" (UID: "10d74042-5cd9-4733-9612-2357a2577f0a"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:53:50 crc kubenswrapper[4998]: I1203 16:53:50.900995 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "10d74042-5cd9-4733-9612-2357a2577f0a" (UID: "10d74042-5cd9-4733-9612-2357a2577f0a"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:53:50 crc kubenswrapper[4998]: I1203 16:53:50.952053 4998 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 16:53:50 crc kubenswrapper[4998]: I1203 16:53:50.952080 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b28wt\" (UniqueName: \"kubernetes.io/projected/10d74042-5cd9-4733-9612-2357a2577f0a-kube-api-access-b28wt\") on node \"crc\" DevicePath \"\"" Dec 03 16:53:50 crc kubenswrapper[4998]: I1203 16:53:50.952091 4998 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 03 16:53:50 crc kubenswrapper[4998]: I1203 16:53:50.952100 4998 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 16:53:50 crc kubenswrapper[4998]: I1203 16:53:50.952108 4998 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 03 16:53:50 crc kubenswrapper[4998]: I1203 16:53:50.952117 4998 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:53:50 crc kubenswrapper[4998]: I1203 16:53:50.952126 4998 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/10d74042-5cd9-4733-9612-2357a2577f0a-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 03 16:53:51 crc kubenswrapper[4998]: I1203 16:53:51.343551 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" event={"ID":"10d74042-5cd9-4733-9612-2357a2577f0a","Type":"ContainerDied","Data":"37fc94e07cb3730df8bcc3621d07fb6983bee732a94fa2da6ac191c5de59a43b"} Dec 03 16:53:51 crc kubenswrapper[4998]: I1203 16:53:51.343645 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37fc94e07cb3730df8bcc3621d07fb6983bee732a94fa2da6ac191c5de59a43b" Dec 03 16:53:51 crc kubenswrapper[4998]: I1203 16:53:51.343719 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2" Dec 03 16:53:58 crc kubenswrapper[4998]: I1203 16:53:58.677917 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:53:58 crc kubenswrapper[4998]: E1203 16:53:58.678802 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:54:13 crc kubenswrapper[4998]: I1203 16:54:13.678017 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:54:13 crc kubenswrapper[4998]: E1203 16:54:13.679073 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:54:26 crc kubenswrapper[4998]: I1203 16:54:26.679229 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:54:26 crc kubenswrapper[4998]: E1203 16:54:26.679981 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.834335 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Dec 03 16:54:29 crc kubenswrapper[4998]: E1203 16:54:29.835351 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10d74042-5cd9-4733-9612-2357a2577f0a" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.835370 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="10d74042-5cd9-4733-9612-2357a2577f0a" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.835635 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="10d74042-5cd9-4733-9612-2357a2577f0a" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.836910 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.839494 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.852651 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.865416 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-config-data-custom\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.865485 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-config-data\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.865590 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-scripts\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.865638 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mxsl\" (UniqueName: \"kubernetes.io/projected/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-kube-api-access-4mxsl\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.865667 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.865714 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.865741 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-dev\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.865800 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-sys\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.865885 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-lib-modules\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.865912 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.866044 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-etc-nvme\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.866072 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.866123 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.866161 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.866186 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-run\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.906930 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-nfs-0"] Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.909122 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.912631 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-nfs-config-data" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.916044 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-nfs-0"] Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.968284 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-config-data-custom\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.968337 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-config-data\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.968358 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-scripts\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.968378 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-config-data-custom\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.968403 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mxsl\" (UniqueName: \"kubernetes.io/projected/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-kube-api-access-4mxsl\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.968420 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.968442 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-var-lib-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.968462 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.968482 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-dev\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.968526 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-dev\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.968548 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.968589 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-sys\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.968650 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-var-locks-brick\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.968682 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-scripts\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.968664 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-sys\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.968730 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-lib-modules\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.968836 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-lib-modules\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.968870 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-etc-nvme\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.968886 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-sys\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.968926 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.968936 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-lib-modules\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.969025 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-etc-iscsi\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.969072 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-config-data\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.969117 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-dev\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.969146 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bzz2\" (UniqueName: \"kubernetes.io/projected/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-kube-api-access-2bzz2\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.969181 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-etc-nvme\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.969206 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.969260 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-etc-machine-id\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.969280 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.969319 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.969378 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.969414 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-run\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.969424 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.969465 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-run\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.969500 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-combined-ca-bundle\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.969541 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-var-locks-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.969569 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-run\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.969774 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.970872 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-nfs-2-0"] Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.970883 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.971096 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-etc-nvme\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.972575 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.974873 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-config-data\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.975091 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-nfs-2-config-data" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.976199 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.979087 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-scripts\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:29 crc kubenswrapper[4998]: I1203 16:54:29.987153 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-config-data-custom\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.002580 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mxsl\" (UniqueName: \"kubernetes.io/projected/0f0ad14b-5286-4d27-8dd1-0a4df70120a6-kube-api-access-4mxsl\") pod \"cinder-backup-0\" (UID: \"0f0ad14b-5286-4d27-8dd1-0a4df70120a6\") " pod="openstack/cinder-backup-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.033332 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-nfs-2-0"] Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072029 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-var-locks-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072335 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-run\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072387 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8bbx\" (UniqueName: \"kubernetes.io/projected/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-kube-api-access-m8bbx\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072162 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-var-locks-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072416 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-config-data-custom\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072433 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-config-data-custom\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072505 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-run\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072567 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-var-lib-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072607 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-var-locks-brick\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072609 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-var-lib-cinder\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072640 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-etc-nvme\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072659 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-var-locks-brick\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072663 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-config-data\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072684 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-scripts\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072721 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-var-locks-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072751 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-lib-modules\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072787 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-etc-machine-id\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072807 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-etc-nvme\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072823 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-sys\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072847 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-combined-ca-bundle\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072869 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-lib-modules\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072877 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-etc-iscsi\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072894 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-dev\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072913 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-config-data\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072937 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-dev\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072953 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-etc-iscsi\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072979 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bzz2\" (UniqueName: \"kubernetes.io/projected/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-kube-api-access-2bzz2\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.072995 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-lib-modules\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.073020 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-var-locks-brick\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.073029 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-etc-iscsi\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.073098 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-etc-nvme\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.073149 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-dev\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.073549 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-sys\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.073649 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-scripts\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.073690 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-etc-machine-id\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.073709 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-run\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.073774 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-sys\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.073797 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-var-lib-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.073835 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-combined-ca-bundle\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.075960 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-etc-machine-id\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.076895 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-config-data-custom\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.077724 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-combined-ca-bundle\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.077729 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-scripts\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.080955 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-config-data\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.094282 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bzz2\" (UniqueName: \"kubernetes.io/projected/f9f63cb2-129d-428c-9a3b-f7646f3c8eb1-kube-api-access-2bzz2\") pod \"cinder-volume-nfs-0\" (UID: \"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1\") " pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.157202 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.175170 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-etc-iscsi\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.175211 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-lib-modules\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.175238 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-var-locks-brick\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.175265 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-scripts\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.175281 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-run\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.175315 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-sys\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.175320 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-etc-iscsi\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.175336 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-var-lib-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.175355 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-lib-modules\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.175398 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-var-lib-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.175439 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8bbx\" (UniqueName: \"kubernetes.io/projected/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-kube-api-access-m8bbx\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.175453 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-var-locks-brick\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.175461 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-config-data-custom\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.175483 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-sys\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.175505 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-run\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.175515 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-etc-nvme\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.175532 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-config-data\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.175550 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-var-locks-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.175581 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-etc-machine-id\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.175608 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-combined-ca-bundle\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.175637 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-dev\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.175696 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-dev\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.175906 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-var-locks-cinder\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.175961 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-etc-nvme\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.176137 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-etc-machine-id\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.180472 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-config-data\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.180960 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-scripts\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.181136 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-config-data-custom\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.182623 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-combined-ca-bundle\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.197610 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8bbx\" (UniqueName: \"kubernetes.io/projected/d83ce9f0-9df6-470e-8dc5-ea647fd082ca-kube-api-access-m8bbx\") pod \"cinder-volume-nfs-2-0\" (UID: \"d83ce9f0-9df6-470e-8dc5-ea647fd082ca\") " pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.226443 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.226646 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.790117 4998 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 16:54:30 crc kubenswrapper[4998]: I1203 16:54:30.799151 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 03 16:54:31 crc kubenswrapper[4998]: I1203 16:54:31.034003 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-nfs-2-0"] Dec 03 16:54:31 crc kubenswrapper[4998]: W1203 16:54:31.035250 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd83ce9f0_9df6_470e_8dc5_ea647fd082ca.slice/crio-d3fb73cf55cb56e4ca5cdbb0671305999b0bd228f26863cb85f11c25fc558a5d WatchSource:0}: Error finding container d3fb73cf55cb56e4ca5cdbb0671305999b0bd228f26863cb85f11c25fc558a5d: Status 404 returned error can't find the container with id d3fb73cf55cb56e4ca5cdbb0671305999b0bd228f26863cb85f11c25fc558a5d Dec 03 16:54:31 crc kubenswrapper[4998]: I1203 16:54:31.150113 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-nfs-0"] Dec 03 16:54:31 crc kubenswrapper[4998]: I1203 16:54:31.783044 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-0" event={"ID":"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1","Type":"ContainerStarted","Data":"3caf64a5d8edb1e05f89d83e6ee1b32eb332ed7e37583ed643ba09accea75cd1"} Dec 03 16:54:31 crc kubenswrapper[4998]: I1203 16:54:31.783518 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-0" event={"ID":"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1","Type":"ContainerStarted","Data":"ee726dd3253c30242022d86e28851dbcaa71ca1a62651497f04aae1c56be44b1"} Dec 03 16:54:31 crc kubenswrapper[4998]: I1203 16:54:31.784871 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-2-0" event={"ID":"d83ce9f0-9df6-470e-8dc5-ea647fd082ca","Type":"ContainerStarted","Data":"a301977c916b98293b7c699b7206d167ee6c4ca7a77b0bd49cc634e89eb3f80f"} Dec 03 16:54:31 crc kubenswrapper[4998]: I1203 16:54:31.784898 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-2-0" event={"ID":"d83ce9f0-9df6-470e-8dc5-ea647fd082ca","Type":"ContainerStarted","Data":"d3fb73cf55cb56e4ca5cdbb0671305999b0bd228f26863cb85f11c25fc558a5d"} Dec 03 16:54:31 crc kubenswrapper[4998]: I1203 16:54:31.787316 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"0f0ad14b-5286-4d27-8dd1-0a4df70120a6","Type":"ContainerStarted","Data":"1ce7f3d79aa9b8012a9977899c612c8fdc98db42f3e57d7b65d76028980836a6"} Dec 03 16:54:31 crc kubenswrapper[4998]: I1203 16:54:31.787343 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"0f0ad14b-5286-4d27-8dd1-0a4df70120a6","Type":"ContainerStarted","Data":"1f8ae6ea2631d847e9f391910479e8e8d43f42851d1500d4e2ca0dd6b22ce497"} Dec 03 16:54:31 crc kubenswrapper[4998]: I1203 16:54:31.787352 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"0f0ad14b-5286-4d27-8dd1-0a4df70120a6","Type":"ContainerStarted","Data":"3b19eecfdf246901294466a348c8eb04de3e06f4bd50da52f129f6032e8e38f4"} Dec 03 16:54:31 crc kubenswrapper[4998]: I1203 16:54:31.822971 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.602432425 podStartE2EDuration="2.822947736s" podCreationTimestamp="2025-12-03 16:54:29 +0000 UTC" firstStartedPulling="2025-12-03 16:54:30.789913863 +0000 UTC m=+3049.401614086" lastFinishedPulling="2025-12-03 16:54:31.010429174 +0000 UTC m=+3049.622129397" observedRunningTime="2025-12-03 16:54:31.816050634 +0000 UTC m=+3050.427750857" watchObservedRunningTime="2025-12-03 16:54:31.822947736 +0000 UTC m=+3050.434647969" Dec 03 16:54:32 crc kubenswrapper[4998]: I1203 16:54:32.797787 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-0" event={"ID":"f9f63cb2-129d-428c-9a3b-f7646f3c8eb1","Type":"ContainerStarted","Data":"fee7a0b452cfe821af63426c57bbb944da46435e3ef486094791433a96118f50"} Dec 03 16:54:32 crc kubenswrapper[4998]: I1203 16:54:32.799404 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-nfs-2-0" event={"ID":"d83ce9f0-9df6-470e-8dc5-ea647fd082ca","Type":"ContainerStarted","Data":"eed7aa4d7a6e16d4f33839d3ffb10956bb17bfe4df58ee1578aec20f92cb8de0"} Dec 03 16:54:32 crc kubenswrapper[4998]: I1203 16:54:32.825960 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-nfs-0" podStartSLOduration=3.757520225 podStartE2EDuration="3.825941909s" podCreationTimestamp="2025-12-03 16:54:29 +0000 UTC" firstStartedPulling="2025-12-03 16:54:31.271677598 +0000 UTC m=+3049.883377821" lastFinishedPulling="2025-12-03 16:54:31.340099272 +0000 UTC m=+3049.951799505" observedRunningTime="2025-12-03 16:54:32.820982518 +0000 UTC m=+3051.432682741" watchObservedRunningTime="2025-12-03 16:54:32.825941909 +0000 UTC m=+3051.437642132" Dec 03 16:54:32 crc kubenswrapper[4998]: I1203 16:54:32.861478 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-nfs-2-0" podStartSLOduration=3.5926891420000002 podStartE2EDuration="3.861452634s" podCreationTimestamp="2025-12-03 16:54:29 +0000 UTC" firstStartedPulling="2025-12-03 16:54:31.03797371 +0000 UTC m=+3049.649673933" lastFinishedPulling="2025-12-03 16:54:31.306737202 +0000 UTC m=+3049.918437425" observedRunningTime="2025-12-03 16:54:32.853916396 +0000 UTC m=+3051.465616619" watchObservedRunningTime="2025-12-03 16:54:32.861452634 +0000 UTC m=+3051.473152857" Dec 03 16:54:35 crc kubenswrapper[4998]: I1203 16:54:35.157349 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Dec 03 16:54:35 crc kubenswrapper[4998]: I1203 16:54:35.227336 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:35 crc kubenswrapper[4998]: I1203 16:54:35.227420 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:40 crc kubenswrapper[4998]: I1203 16:54:40.314162 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Dec 03 16:54:40 crc kubenswrapper[4998]: I1203 16:54:40.373641 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-nfs-0" Dec 03 16:54:40 crc kubenswrapper[4998]: I1203 16:54:40.596652 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-nfs-2-0" Dec 03 16:54:41 crc kubenswrapper[4998]: I1203 16:54:41.692125 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:54:41 crc kubenswrapper[4998]: E1203 16:54:41.692839 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:54:55 crc kubenswrapper[4998]: I1203 16:54:55.679743 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:54:55 crc kubenswrapper[4998]: E1203 16:54:55.681417 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:55:06 crc kubenswrapper[4998]: I1203 16:55:06.677813 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:55:06 crc kubenswrapper[4998]: E1203 16:55:06.678585 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:55:21 crc kubenswrapper[4998]: I1203 16:55:21.686044 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:55:21 crc kubenswrapper[4998]: E1203 16:55:21.687039 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:55:32 crc kubenswrapper[4998]: I1203 16:55:32.679015 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:55:32 crc kubenswrapper[4998]: E1203 16:55:32.679840 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:55:33 crc kubenswrapper[4998]: I1203 16:55:33.457432 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 16:55:33 crc kubenswrapper[4998]: I1203 16:55:33.457778 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="a70537b1-93bc-48c3-9bef-fd9030158cad" containerName="prometheus" containerID="cri-o://a7bc207f9aabc3bd16ffd9cad967412d25c52c69865f816487cd0b58b6c4d55b" gracePeriod=600 Dec 03 16:55:33 crc kubenswrapper[4998]: I1203 16:55:33.457859 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="a70537b1-93bc-48c3-9bef-fd9030158cad" containerName="config-reloader" containerID="cri-o://21be89612a3b0e4cd79f2ace91e7ce2132daa6eb2b02c74b17c740d1d544d851" gracePeriod=600 Dec 03 16:55:33 crc kubenswrapper[4998]: I1203 16:55:33.457842 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="a70537b1-93bc-48c3-9bef-fd9030158cad" containerName="thanos-sidecar" containerID="cri-o://d6a4eaeddd7b46a9b6c18edbb3bab49896a9ae66616e1863c624492c605f1345" gracePeriod=600 Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.521279 4998 generic.go:334] "Generic (PLEG): container finished" podID="a70537b1-93bc-48c3-9bef-fd9030158cad" containerID="d6a4eaeddd7b46a9b6c18edbb3bab49896a9ae66616e1863c624492c605f1345" exitCode=0 Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.521676 4998 generic.go:334] "Generic (PLEG): container finished" podID="a70537b1-93bc-48c3-9bef-fd9030158cad" containerID="21be89612a3b0e4cd79f2ace91e7ce2132daa6eb2b02c74b17c740d1d544d851" exitCode=0 Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.521686 4998 generic.go:334] "Generic (PLEG): container finished" podID="a70537b1-93bc-48c3-9bef-fd9030158cad" containerID="a7bc207f9aabc3bd16ffd9cad967412d25c52c69865f816487cd0b58b6c4d55b" exitCode=0 Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.521372 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a70537b1-93bc-48c3-9bef-fd9030158cad","Type":"ContainerDied","Data":"d6a4eaeddd7b46a9b6c18edbb3bab49896a9ae66616e1863c624492c605f1345"} Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.521719 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a70537b1-93bc-48c3-9bef-fd9030158cad","Type":"ContainerDied","Data":"21be89612a3b0e4cd79f2ace91e7ce2132daa6eb2b02c74b17c740d1d544d851"} Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.521743 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a70537b1-93bc-48c3-9bef-fd9030158cad","Type":"ContainerDied","Data":"a7bc207f9aabc3bd16ffd9cad967412d25c52c69865f816487cd0b58b6c4d55b"} Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.521769 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"a70537b1-93bc-48c3-9bef-fd9030158cad","Type":"ContainerDied","Data":"42cc47506f3d7b982ad83a6d55f67124b3d014b607d4dc73e61b9a42e1f91d65"} Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.521779 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42cc47506f3d7b982ad83a6d55f67124b3d014b607d4dc73e61b9a42e1f91d65" Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.529470 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.604137 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-thanos-prometheus-http-client-file\") pod \"a70537b1-93bc-48c3-9bef-fd9030158cad\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.604238 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-secret-combined-ca-bundle\") pod \"a70537b1-93bc-48c3-9bef-fd9030158cad\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.604286 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bh8n\" (UniqueName: \"kubernetes.io/projected/a70537b1-93bc-48c3-9bef-fd9030158cad-kube-api-access-6bh8n\") pod \"a70537b1-93bc-48c3-9bef-fd9030158cad\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.604321 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"a70537b1-93bc-48c3-9bef-fd9030158cad\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.604428 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c9889e3e-2478-469f-b92d-e203678c4927\") pod \"a70537b1-93bc-48c3-9bef-fd9030158cad\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.604471 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a70537b1-93bc-48c3-9bef-fd9030158cad-tls-assets\") pod \"a70537b1-93bc-48c3-9bef-fd9030158cad\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.604505 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/a70537b1-93bc-48c3-9bef-fd9030158cad-prometheus-metric-storage-rulefiles-0\") pod \"a70537b1-93bc-48c3-9bef-fd9030158cad\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.604534 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-config\") pod \"a70537b1-93bc-48c3-9bef-fd9030158cad\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.604605 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"a70537b1-93bc-48c3-9bef-fd9030158cad\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.604657 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a70537b1-93bc-48c3-9bef-fd9030158cad-config-out\") pod \"a70537b1-93bc-48c3-9bef-fd9030158cad\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.604725 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-web-config\") pod \"a70537b1-93bc-48c3-9bef-fd9030158cad\" (UID: \"a70537b1-93bc-48c3-9bef-fd9030158cad\") " Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.606308 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a70537b1-93bc-48c3-9bef-fd9030158cad-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "a70537b1-93bc-48c3-9bef-fd9030158cad" (UID: "a70537b1-93bc-48c3-9bef-fd9030158cad"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.613508 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "a70537b1-93bc-48c3-9bef-fd9030158cad" (UID: "a70537b1-93bc-48c3-9bef-fd9030158cad"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.613547 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-config" (OuterVolumeSpecName: "config") pod "a70537b1-93bc-48c3-9bef-fd9030158cad" (UID: "a70537b1-93bc-48c3-9bef-fd9030158cad"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.613616 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-secret-combined-ca-bundle" (OuterVolumeSpecName: "secret-combined-ca-bundle") pod "a70537b1-93bc-48c3-9bef-fd9030158cad" (UID: "a70537b1-93bc-48c3-9bef-fd9030158cad"). InnerVolumeSpecName "secret-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.613635 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a70537b1-93bc-48c3-9bef-fd9030158cad-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "a70537b1-93bc-48c3-9bef-fd9030158cad" (UID: "a70537b1-93bc-48c3-9bef-fd9030158cad"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.613871 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a70537b1-93bc-48c3-9bef-fd9030158cad-config-out" (OuterVolumeSpecName: "config-out") pod "a70537b1-93bc-48c3-9bef-fd9030158cad" (UID: "a70537b1-93bc-48c3-9bef-fd9030158cad"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.614860 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d") pod "a70537b1-93bc-48c3-9bef-fd9030158cad" (UID: "a70537b1-93bc-48c3-9bef-fd9030158cad"). InnerVolumeSpecName "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.626988 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a70537b1-93bc-48c3-9bef-fd9030158cad-kube-api-access-6bh8n" (OuterVolumeSpecName: "kube-api-access-6bh8n") pod "a70537b1-93bc-48c3-9bef-fd9030158cad" (UID: "a70537b1-93bc-48c3-9bef-fd9030158cad"). InnerVolumeSpecName "kube-api-access-6bh8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.641987 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d") pod "a70537b1-93bc-48c3-9bef-fd9030158cad" (UID: "a70537b1-93bc-48c3-9bef-fd9030158cad"). InnerVolumeSpecName "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.654242 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c9889e3e-2478-469f-b92d-e203678c4927" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "a70537b1-93bc-48c3-9bef-fd9030158cad" (UID: "a70537b1-93bc-48c3-9bef-fd9030158cad"). InnerVolumeSpecName "pvc-c9889e3e-2478-469f-b92d-e203678c4927". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.707954 4998 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a70537b1-93bc-48c3-9bef-fd9030158cad-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.708004 4998 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/a70537b1-93bc-48c3-9bef-fd9030158cad-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.708014 4998 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.708024 4998 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") on node \"crc\" DevicePath \"\"" Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.708033 4998 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a70537b1-93bc-48c3-9bef-fd9030158cad-config-out\") on node \"crc\" DevicePath \"\"" Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.708043 4998 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.708053 4998 reconciler_common.go:293] "Volume detached for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-secret-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.708063 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bh8n\" (UniqueName: \"kubernetes.io/projected/a70537b1-93bc-48c3-9bef-fd9030158cad-kube-api-access-6bh8n\") on node \"crc\" DevicePath \"\"" Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.708072 4998 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") on node \"crc\" DevicePath \"\"" Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.708092 4998 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-c9889e3e-2478-469f-b92d-e203678c4927\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c9889e3e-2478-469f-b92d-e203678c4927\") on node \"crc\" " Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.710629 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-web-config" (OuterVolumeSpecName: "web-config") pod "a70537b1-93bc-48c3-9bef-fd9030158cad" (UID: "a70537b1-93bc-48c3-9bef-fd9030158cad"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.738380 4998 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.739434 4998 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-c9889e3e-2478-469f-b92d-e203678c4927" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c9889e3e-2478-469f-b92d-e203678c4927") on node "crc" Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.810441 4998 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a70537b1-93bc-48c3-9bef-fd9030158cad-web-config\") on node \"crc\" DevicePath \"\"" Dec 03 16:55:34 crc kubenswrapper[4998]: I1203 16:55:34.810484 4998 reconciler_common.go:293] "Volume detached for volume \"pvc-c9889e3e-2478-469f-b92d-e203678c4927\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c9889e3e-2478-469f-b92d-e203678c4927\") on node \"crc\" DevicePath \"\"" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.529460 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.566281 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.576539 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.589241 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 16:55:35 crc kubenswrapper[4998]: E1203 16:55:35.589653 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a70537b1-93bc-48c3-9bef-fd9030158cad" containerName="prometheus" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.589670 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="a70537b1-93bc-48c3-9bef-fd9030158cad" containerName="prometheus" Dec 03 16:55:35 crc kubenswrapper[4998]: E1203 16:55:35.589695 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a70537b1-93bc-48c3-9bef-fd9030158cad" containerName="thanos-sidecar" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.589701 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="a70537b1-93bc-48c3-9bef-fd9030158cad" containerName="thanos-sidecar" Dec 03 16:55:35 crc kubenswrapper[4998]: E1203 16:55:35.589717 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a70537b1-93bc-48c3-9bef-fd9030158cad" containerName="init-config-reloader" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.589723 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="a70537b1-93bc-48c3-9bef-fd9030158cad" containerName="init-config-reloader" Dec 03 16:55:35 crc kubenswrapper[4998]: E1203 16:55:35.589738 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a70537b1-93bc-48c3-9bef-fd9030158cad" containerName="config-reloader" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.589745 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="a70537b1-93bc-48c3-9bef-fd9030158cad" containerName="config-reloader" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.589949 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="a70537b1-93bc-48c3-9bef-fd9030158cad" containerName="config-reloader" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.589961 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="a70537b1-93bc-48c3-9bef-fd9030158cad" containerName="prometheus" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.589971 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="a70537b1-93bc-48c3-9bef-fd9030158cad" containerName="thanos-sidecar" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.591652 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.594511 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-fx7rj" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.594798 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.594749 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.595197 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.596416 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.602509 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.627098 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.627149 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.627328 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.627439 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.627522 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz8dc\" (UniqueName: \"kubernetes.io/projected/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-kube-api-access-dz8dc\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.627581 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.627618 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c9889e3e-2478-469f-b92d-e203678c4927\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c9889e3e-2478-469f-b92d-e203678c4927\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.627660 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.627799 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.627863 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.627976 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-config\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.689195 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a70537b1-93bc-48c3-9bef-fd9030158cad" path="/var/lib/kubelet/pods/a70537b1-93bc-48c3-9bef-fd9030158cad/volumes" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.692319 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.731106 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.731164 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.731223 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-config\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.731263 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.731286 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.731347 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.731432 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.731477 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz8dc\" (UniqueName: \"kubernetes.io/projected/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-kube-api-access-dz8dc\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.731523 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.731584 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c9889e3e-2478-469f-b92d-e203678c4927\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c9889e3e-2478-469f-b92d-e203678c4927\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.731622 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.732142 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.745588 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.753021 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.754568 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.754697 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.756309 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.765711 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz8dc\" (UniqueName: \"kubernetes.io/projected/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-kube-api-access-dz8dc\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.768647 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-config\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.769158 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.784315 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.790355 4998 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.790409 4998 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c9889e3e-2478-469f-b92d-e203678c4927\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c9889e3e-2478-469f-b92d-e203678c4927\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/afadf41a682809fbeb8f0a8dc0e25f598b7474e7bc0a2ee0471a59f3f2d1c942/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:35 crc kubenswrapper[4998]: I1203 16:55:35.966388 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c9889e3e-2478-469f-b92d-e203678c4927\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c9889e3e-2478-469f-b92d-e203678c4927\") pod \"prometheus-metric-storage-0\" (UID: \"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1\") " pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:36 crc kubenswrapper[4998]: I1203 16:55:36.210175 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:36 crc kubenswrapper[4998]: I1203 16:55:36.728957 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 16:55:37 crc kubenswrapper[4998]: I1203 16:55:37.553786 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1","Type":"ContainerStarted","Data":"190c8708ee3215d4bd6f22a8ee6bb66b92428c9c9687aa9caa68ab85fe128c47"} Dec 03 16:55:41 crc kubenswrapper[4998]: I1203 16:55:41.600093 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1","Type":"ContainerStarted","Data":"a2110ba8ffe223620ce17a3d468bb839de67ff0e98a53d6166dccc6702cb1964"} Dec 03 16:55:43 crc kubenswrapper[4998]: I1203 16:55:43.678744 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:55:43 crc kubenswrapper[4998]: E1203 16:55:43.679650 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:55:50 crc kubenswrapper[4998]: I1203 16:55:50.737999 4998 generic.go:334] "Generic (PLEG): container finished" podID="b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1" containerID="a2110ba8ffe223620ce17a3d468bb839de67ff0e98a53d6166dccc6702cb1964" exitCode=0 Dec 03 16:55:50 crc kubenswrapper[4998]: I1203 16:55:50.738110 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1","Type":"ContainerDied","Data":"a2110ba8ffe223620ce17a3d468bb839de67ff0e98a53d6166dccc6702cb1964"} Dec 03 16:55:51 crc kubenswrapper[4998]: I1203 16:55:51.757355 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1","Type":"ContainerStarted","Data":"3ca0bc37ae07e28da27eda61ebe3b8c7cc4b6463cb4617d9a1132080742acbb4"} Dec 03 16:55:55 crc kubenswrapper[4998]: I1203 16:55:55.805440 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1","Type":"ContainerStarted","Data":"f3e5d3754a3f6735eadc94de145cbd610ca10c0761a6fd9f51de65efd4b09df4"} Dec 03 16:55:55 crc kubenswrapper[4998]: I1203 16:55:55.805984 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1","Type":"ContainerStarted","Data":"58b89fc911b30fe2f88567677ff2fd7efb18b431f80b8cdb4d1e771468124450"} Dec 03 16:55:55 crc kubenswrapper[4998]: I1203 16:55:55.847365 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=20.847336627 podStartE2EDuration="20.847336627s" podCreationTimestamp="2025-12-03 16:55:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 16:55:55.835604818 +0000 UTC m=+3134.447305051" watchObservedRunningTime="2025-12-03 16:55:55.847336627 +0000 UTC m=+3134.459036890" Dec 03 16:55:56 crc kubenswrapper[4998]: I1203 16:55:56.210570 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 03 16:55:56 crc kubenswrapper[4998]: I1203 16:55:56.677832 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:55:56 crc kubenswrapper[4998]: E1203 16:55:56.678345 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:56:01 crc kubenswrapper[4998]: I1203 16:56:01.024819 4998 scope.go:117] "RemoveContainer" containerID="d6a4eaeddd7b46a9b6c18edbb3bab49896a9ae66616e1863c624492c605f1345" Dec 03 16:56:01 crc kubenswrapper[4998]: I1203 16:56:01.051967 4998 scope.go:117] "RemoveContainer" containerID="9bd7f454c517ad30a0d2fb22370f32404af1a46c10a6bbd97476e74751b2c2c1" Dec 03 16:56:01 crc kubenswrapper[4998]: I1203 16:56:01.078379 4998 scope.go:117] "RemoveContainer" containerID="a7bc207f9aabc3bd16ffd9cad967412d25c52c69865f816487cd0b58b6c4d55b" Dec 03 16:56:01 crc kubenswrapper[4998]: I1203 16:56:01.160892 4998 scope.go:117] "RemoveContainer" containerID="21be89612a3b0e4cd79f2ace91e7ce2132daa6eb2b02c74b17c740d1d544d851" Dec 03 16:56:06 crc kubenswrapper[4998]: I1203 16:56:06.210912 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 03 16:56:06 crc kubenswrapper[4998]: I1203 16:56:06.219472 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 03 16:56:06 crc kubenswrapper[4998]: I1203 16:56:06.941830 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 03 16:56:07 crc kubenswrapper[4998]: I1203 16:56:07.678412 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:56:07 crc kubenswrapper[4998]: E1203 16:56:07.679161 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:56:19 crc kubenswrapper[4998]: I1203 16:56:19.678718 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:56:19 crc kubenswrapper[4998]: E1203 16:56:19.680072 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 16:56:26 crc kubenswrapper[4998]: I1203 16:56:26.954855 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 16:56:26 crc kubenswrapper[4998]: I1203 16:56:26.957748 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 16:56:26 crc kubenswrapper[4998]: I1203 16:56:26.960596 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 03 16:56:26 crc kubenswrapper[4998]: I1203 16:56:26.962057 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 03 16:56:26 crc kubenswrapper[4998]: I1203 16:56:26.962194 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-28bzn" Dec 03 16:56:26 crc kubenswrapper[4998]: I1203 16:56:26.964178 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 03 16:56:26 crc kubenswrapper[4998]: I1203 16:56:26.975748 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.049428 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.049493 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.049523 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.049634 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-config-data\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.049663 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.049914 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.050700 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ww9s\" (UniqueName: \"kubernetes.io/projected/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-kube-api-access-8ww9s\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.051276 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.051433 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.156744 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.157358 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.157461 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.157508 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.157588 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.157701 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-config-data\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.157730 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.157772 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.157836 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ww9s\" (UniqueName: \"kubernetes.io/projected/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-kube-api-access-8ww9s\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.160751 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.169115 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.169690 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.170022 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.170392 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.171163 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-config-data\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.176958 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.194946 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ww9s\" (UniqueName: \"kubernetes.io/projected/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-kube-api-access-8ww9s\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.195306 4998 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.222500 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.285863 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 16:56:27 crc kubenswrapper[4998]: I1203 16:56:27.984695 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 16:56:28 crc kubenswrapper[4998]: I1203 16:56:28.175957 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"27edc0ce-edf8-4c0e-bd5f-79e58650eed1","Type":"ContainerStarted","Data":"63a2fe08a9bdff7f5ed4f9daa8d5547870580605d42e0286337953d2a8d6b65b"} Dec 03 16:56:32 crc kubenswrapper[4998]: I1203 16:56:32.677661 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:56:39 crc kubenswrapper[4998]: I1203 16:56:39.286490 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"84aefa566144217de53fa4d3aefaa1bc2dcd21fcd378c021e19510c75d4d205e"} Dec 03 16:56:42 crc kubenswrapper[4998]: I1203 16:56:42.323993 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"27edc0ce-edf8-4c0e-bd5f-79e58650eed1","Type":"ContainerStarted","Data":"55cea1c7702ffa07417318378686a0b88bb4d847c19bf491fc95f56a149afbc4"} Dec 03 16:56:42 crc kubenswrapper[4998]: I1203 16:56:42.352976 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=5.217707876 podStartE2EDuration="17.35295211s" podCreationTimestamp="2025-12-03 16:56:25 +0000 UTC" firstStartedPulling="2025-12-03 16:56:27.987689141 +0000 UTC m=+3166.599389364" lastFinishedPulling="2025-12-03 16:56:40.122933365 +0000 UTC m=+3178.734633598" observedRunningTime="2025-12-03 16:56:42.344820048 +0000 UTC m=+3180.956520331" watchObservedRunningTime="2025-12-03 16:56:42.35295211 +0000 UTC m=+3180.964652343" Dec 03 16:56:56 crc kubenswrapper[4998]: I1203 16:56:56.927801 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rzsmv"] Dec 03 16:56:56 crc kubenswrapper[4998]: I1203 16:56:56.930477 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rzsmv" Dec 03 16:56:56 crc kubenswrapper[4998]: I1203 16:56:56.949382 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rzsmv"] Dec 03 16:56:57 crc kubenswrapper[4998]: I1203 16:56:57.040871 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a6110f-3034-45bd-89cc-176b1a7a2cb7-utilities\") pod \"redhat-marketplace-rzsmv\" (UID: \"57a6110f-3034-45bd-89cc-176b1a7a2cb7\") " pod="openshift-marketplace/redhat-marketplace-rzsmv" Dec 03 16:56:57 crc kubenswrapper[4998]: I1203 16:56:57.040932 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmxkw\" (UniqueName: \"kubernetes.io/projected/57a6110f-3034-45bd-89cc-176b1a7a2cb7-kube-api-access-nmxkw\") pod \"redhat-marketplace-rzsmv\" (UID: \"57a6110f-3034-45bd-89cc-176b1a7a2cb7\") " pod="openshift-marketplace/redhat-marketplace-rzsmv" Dec 03 16:56:57 crc kubenswrapper[4998]: I1203 16:56:57.041167 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a6110f-3034-45bd-89cc-176b1a7a2cb7-catalog-content\") pod \"redhat-marketplace-rzsmv\" (UID: \"57a6110f-3034-45bd-89cc-176b1a7a2cb7\") " pod="openshift-marketplace/redhat-marketplace-rzsmv" Dec 03 16:56:57 crc kubenswrapper[4998]: I1203 16:56:57.143637 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a6110f-3034-45bd-89cc-176b1a7a2cb7-utilities\") pod \"redhat-marketplace-rzsmv\" (UID: \"57a6110f-3034-45bd-89cc-176b1a7a2cb7\") " pod="openshift-marketplace/redhat-marketplace-rzsmv" Dec 03 16:56:57 crc kubenswrapper[4998]: I1203 16:56:57.143713 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmxkw\" (UniqueName: \"kubernetes.io/projected/57a6110f-3034-45bd-89cc-176b1a7a2cb7-kube-api-access-nmxkw\") pod \"redhat-marketplace-rzsmv\" (UID: \"57a6110f-3034-45bd-89cc-176b1a7a2cb7\") " pod="openshift-marketplace/redhat-marketplace-rzsmv" Dec 03 16:56:57 crc kubenswrapper[4998]: I1203 16:56:57.143862 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a6110f-3034-45bd-89cc-176b1a7a2cb7-catalog-content\") pod \"redhat-marketplace-rzsmv\" (UID: \"57a6110f-3034-45bd-89cc-176b1a7a2cb7\") " pod="openshift-marketplace/redhat-marketplace-rzsmv" Dec 03 16:56:57 crc kubenswrapper[4998]: I1203 16:56:57.144311 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a6110f-3034-45bd-89cc-176b1a7a2cb7-catalog-content\") pod \"redhat-marketplace-rzsmv\" (UID: \"57a6110f-3034-45bd-89cc-176b1a7a2cb7\") " pod="openshift-marketplace/redhat-marketplace-rzsmv" Dec 03 16:56:57 crc kubenswrapper[4998]: I1203 16:56:57.144916 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a6110f-3034-45bd-89cc-176b1a7a2cb7-utilities\") pod \"redhat-marketplace-rzsmv\" (UID: \"57a6110f-3034-45bd-89cc-176b1a7a2cb7\") " pod="openshift-marketplace/redhat-marketplace-rzsmv" Dec 03 16:56:57 crc kubenswrapper[4998]: I1203 16:56:57.176389 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmxkw\" (UniqueName: \"kubernetes.io/projected/57a6110f-3034-45bd-89cc-176b1a7a2cb7-kube-api-access-nmxkw\") pod \"redhat-marketplace-rzsmv\" (UID: \"57a6110f-3034-45bd-89cc-176b1a7a2cb7\") " pod="openshift-marketplace/redhat-marketplace-rzsmv" Dec 03 16:56:57 crc kubenswrapper[4998]: I1203 16:56:57.262619 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rzsmv" Dec 03 16:56:57 crc kubenswrapper[4998]: I1203 16:56:57.908664 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rzsmv"] Dec 03 16:56:57 crc kubenswrapper[4998]: W1203 16:56:57.915887 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57a6110f_3034_45bd_89cc_176b1a7a2cb7.slice/crio-9a7b433c95bbc66b1ed05e77113a84f19dcdaaab94723e54adbe293fb2a46431 WatchSource:0}: Error finding container 9a7b433c95bbc66b1ed05e77113a84f19dcdaaab94723e54adbe293fb2a46431: Status 404 returned error can't find the container with id 9a7b433c95bbc66b1ed05e77113a84f19dcdaaab94723e54adbe293fb2a46431 Dec 03 16:56:57 crc kubenswrapper[4998]: I1203 16:56:57.931831 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rzsmv" event={"ID":"57a6110f-3034-45bd-89cc-176b1a7a2cb7","Type":"ContainerStarted","Data":"9a7b433c95bbc66b1ed05e77113a84f19dcdaaab94723e54adbe293fb2a46431"} Dec 03 16:56:58 crc kubenswrapper[4998]: I1203 16:56:58.944045 4998 generic.go:334] "Generic (PLEG): container finished" podID="57a6110f-3034-45bd-89cc-176b1a7a2cb7" containerID="374464700e2b9db2be22bbb601cba0881bf1cbf484fbf47de550ab712a2fdb6b" exitCode=0 Dec 03 16:56:58 crc kubenswrapper[4998]: I1203 16:56:58.946634 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rzsmv" event={"ID":"57a6110f-3034-45bd-89cc-176b1a7a2cb7","Type":"ContainerDied","Data":"374464700e2b9db2be22bbb601cba0881bf1cbf484fbf47de550ab712a2fdb6b"} Dec 03 16:56:59 crc kubenswrapper[4998]: I1203 16:56:59.957081 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rzsmv" event={"ID":"57a6110f-3034-45bd-89cc-176b1a7a2cb7","Type":"ContainerStarted","Data":"8559afd83ea25472a63f71fc7f27f0b441750ec4189abdcab6aaaab99c5c8dff"} Dec 03 16:57:00 crc kubenswrapper[4998]: I1203 16:57:00.969928 4998 generic.go:334] "Generic (PLEG): container finished" podID="57a6110f-3034-45bd-89cc-176b1a7a2cb7" containerID="8559afd83ea25472a63f71fc7f27f0b441750ec4189abdcab6aaaab99c5c8dff" exitCode=0 Dec 03 16:57:00 crc kubenswrapper[4998]: I1203 16:57:00.970036 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rzsmv" event={"ID":"57a6110f-3034-45bd-89cc-176b1a7a2cb7","Type":"ContainerDied","Data":"8559afd83ea25472a63f71fc7f27f0b441750ec4189abdcab6aaaab99c5c8dff"} Dec 03 16:57:01 crc kubenswrapper[4998]: I1203 16:57:01.988037 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rzsmv" event={"ID":"57a6110f-3034-45bd-89cc-176b1a7a2cb7","Type":"ContainerStarted","Data":"3be23782e10beb64b05c264edb08fb65b1fadd2c717e968d79f86cacdf5532ab"} Dec 03 16:57:02 crc kubenswrapper[4998]: I1203 16:57:02.023671 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rzsmv" podStartSLOduration=3.56779251 podStartE2EDuration="6.023651252s" podCreationTimestamp="2025-12-03 16:56:56 +0000 UTC" firstStartedPulling="2025-12-03 16:56:58.947227469 +0000 UTC m=+3197.558927702" lastFinishedPulling="2025-12-03 16:57:01.403086191 +0000 UTC m=+3200.014786444" observedRunningTime="2025-12-03 16:57:02.008367054 +0000 UTC m=+3200.620067277" watchObservedRunningTime="2025-12-03 16:57:02.023651252 +0000 UTC m=+3200.635351475" Dec 03 16:57:07 crc kubenswrapper[4998]: I1203 16:57:07.262860 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rzsmv" Dec 03 16:57:07 crc kubenswrapper[4998]: I1203 16:57:07.263548 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rzsmv" Dec 03 16:57:07 crc kubenswrapper[4998]: I1203 16:57:07.349411 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rzsmv" Dec 03 16:57:08 crc kubenswrapper[4998]: I1203 16:57:08.138845 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rzsmv" Dec 03 16:57:08 crc kubenswrapper[4998]: I1203 16:57:08.196559 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rzsmv"] Dec 03 16:57:10 crc kubenswrapper[4998]: I1203 16:57:10.071923 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rzsmv" podUID="57a6110f-3034-45bd-89cc-176b1a7a2cb7" containerName="registry-server" containerID="cri-o://3be23782e10beb64b05c264edb08fb65b1fadd2c717e968d79f86cacdf5532ab" gracePeriod=2 Dec 03 16:57:10 crc kubenswrapper[4998]: I1203 16:57:10.577999 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rzsmv" Dec 03 16:57:10 crc kubenswrapper[4998]: I1203 16:57:10.643281 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a6110f-3034-45bd-89cc-176b1a7a2cb7-utilities\") pod \"57a6110f-3034-45bd-89cc-176b1a7a2cb7\" (UID: \"57a6110f-3034-45bd-89cc-176b1a7a2cb7\") " Dec 03 16:57:10 crc kubenswrapper[4998]: I1203 16:57:10.643387 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmxkw\" (UniqueName: \"kubernetes.io/projected/57a6110f-3034-45bd-89cc-176b1a7a2cb7-kube-api-access-nmxkw\") pod \"57a6110f-3034-45bd-89cc-176b1a7a2cb7\" (UID: \"57a6110f-3034-45bd-89cc-176b1a7a2cb7\") " Dec 03 16:57:10 crc kubenswrapper[4998]: I1203 16:57:10.643447 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a6110f-3034-45bd-89cc-176b1a7a2cb7-catalog-content\") pod \"57a6110f-3034-45bd-89cc-176b1a7a2cb7\" (UID: \"57a6110f-3034-45bd-89cc-176b1a7a2cb7\") " Dec 03 16:57:10 crc kubenswrapper[4998]: I1203 16:57:10.645413 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a6110f-3034-45bd-89cc-176b1a7a2cb7-utilities" (OuterVolumeSpecName: "utilities") pod "57a6110f-3034-45bd-89cc-176b1a7a2cb7" (UID: "57a6110f-3034-45bd-89cc-176b1a7a2cb7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:57:10 crc kubenswrapper[4998]: I1203 16:57:10.653650 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a6110f-3034-45bd-89cc-176b1a7a2cb7-kube-api-access-nmxkw" (OuterVolumeSpecName: "kube-api-access-nmxkw") pod "57a6110f-3034-45bd-89cc-176b1a7a2cb7" (UID: "57a6110f-3034-45bd-89cc-176b1a7a2cb7"). InnerVolumeSpecName "kube-api-access-nmxkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:57:10 crc kubenswrapper[4998]: I1203 16:57:10.683124 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a6110f-3034-45bd-89cc-176b1a7a2cb7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a6110f-3034-45bd-89cc-176b1a7a2cb7" (UID: "57a6110f-3034-45bd-89cc-176b1a7a2cb7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:57:10 crc kubenswrapper[4998]: I1203 16:57:10.746328 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a6110f-3034-45bd-89cc-176b1a7a2cb7-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:57:10 crc kubenswrapper[4998]: I1203 16:57:10.746369 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmxkw\" (UniqueName: \"kubernetes.io/projected/57a6110f-3034-45bd-89cc-176b1a7a2cb7-kube-api-access-nmxkw\") on node \"crc\" DevicePath \"\"" Dec 03 16:57:10 crc kubenswrapper[4998]: I1203 16:57:10.746381 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a6110f-3034-45bd-89cc-176b1a7a2cb7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:57:11 crc kubenswrapper[4998]: I1203 16:57:11.085847 4998 generic.go:334] "Generic (PLEG): container finished" podID="57a6110f-3034-45bd-89cc-176b1a7a2cb7" containerID="3be23782e10beb64b05c264edb08fb65b1fadd2c717e968d79f86cacdf5532ab" exitCode=0 Dec 03 16:57:11 crc kubenswrapper[4998]: I1203 16:57:11.085897 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rzsmv" event={"ID":"57a6110f-3034-45bd-89cc-176b1a7a2cb7","Type":"ContainerDied","Data":"3be23782e10beb64b05c264edb08fb65b1fadd2c717e968d79f86cacdf5532ab"} Dec 03 16:57:11 crc kubenswrapper[4998]: I1203 16:57:11.085922 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rzsmv" event={"ID":"57a6110f-3034-45bd-89cc-176b1a7a2cb7","Type":"ContainerDied","Data":"9a7b433c95bbc66b1ed05e77113a84f19dcdaaab94723e54adbe293fb2a46431"} Dec 03 16:57:11 crc kubenswrapper[4998]: I1203 16:57:11.085940 4998 scope.go:117] "RemoveContainer" containerID="3be23782e10beb64b05c264edb08fb65b1fadd2c717e968d79f86cacdf5532ab" Dec 03 16:57:11 crc kubenswrapper[4998]: I1203 16:57:11.086081 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rzsmv" Dec 03 16:57:11 crc kubenswrapper[4998]: I1203 16:57:11.116140 4998 scope.go:117] "RemoveContainer" containerID="8559afd83ea25472a63f71fc7f27f0b441750ec4189abdcab6aaaab99c5c8dff" Dec 03 16:57:11 crc kubenswrapper[4998]: I1203 16:57:11.149410 4998 scope.go:117] "RemoveContainer" containerID="374464700e2b9db2be22bbb601cba0881bf1cbf484fbf47de550ab712a2fdb6b" Dec 03 16:57:11 crc kubenswrapper[4998]: I1203 16:57:11.154920 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rzsmv"] Dec 03 16:57:11 crc kubenswrapper[4998]: I1203 16:57:11.171541 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rzsmv"] Dec 03 16:57:11 crc kubenswrapper[4998]: I1203 16:57:11.234389 4998 scope.go:117] "RemoveContainer" containerID="3be23782e10beb64b05c264edb08fb65b1fadd2c717e968d79f86cacdf5532ab" Dec 03 16:57:11 crc kubenswrapper[4998]: E1203 16:57:11.237199 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3be23782e10beb64b05c264edb08fb65b1fadd2c717e968d79f86cacdf5532ab\": container with ID starting with 3be23782e10beb64b05c264edb08fb65b1fadd2c717e968d79f86cacdf5532ab not found: ID does not exist" containerID="3be23782e10beb64b05c264edb08fb65b1fadd2c717e968d79f86cacdf5532ab" Dec 03 16:57:11 crc kubenswrapper[4998]: I1203 16:57:11.237533 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3be23782e10beb64b05c264edb08fb65b1fadd2c717e968d79f86cacdf5532ab"} err="failed to get container status \"3be23782e10beb64b05c264edb08fb65b1fadd2c717e968d79f86cacdf5532ab\": rpc error: code = NotFound desc = could not find container \"3be23782e10beb64b05c264edb08fb65b1fadd2c717e968d79f86cacdf5532ab\": container with ID starting with 3be23782e10beb64b05c264edb08fb65b1fadd2c717e968d79f86cacdf5532ab not found: ID does not exist" Dec 03 16:57:11 crc kubenswrapper[4998]: I1203 16:57:11.237560 4998 scope.go:117] "RemoveContainer" containerID="8559afd83ea25472a63f71fc7f27f0b441750ec4189abdcab6aaaab99c5c8dff" Dec 03 16:57:11 crc kubenswrapper[4998]: E1203 16:57:11.239644 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8559afd83ea25472a63f71fc7f27f0b441750ec4189abdcab6aaaab99c5c8dff\": container with ID starting with 8559afd83ea25472a63f71fc7f27f0b441750ec4189abdcab6aaaab99c5c8dff not found: ID does not exist" containerID="8559afd83ea25472a63f71fc7f27f0b441750ec4189abdcab6aaaab99c5c8dff" Dec 03 16:57:11 crc kubenswrapper[4998]: I1203 16:57:11.239734 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8559afd83ea25472a63f71fc7f27f0b441750ec4189abdcab6aaaab99c5c8dff"} err="failed to get container status \"8559afd83ea25472a63f71fc7f27f0b441750ec4189abdcab6aaaab99c5c8dff\": rpc error: code = NotFound desc = could not find container \"8559afd83ea25472a63f71fc7f27f0b441750ec4189abdcab6aaaab99c5c8dff\": container with ID starting with 8559afd83ea25472a63f71fc7f27f0b441750ec4189abdcab6aaaab99c5c8dff not found: ID does not exist" Dec 03 16:57:11 crc kubenswrapper[4998]: I1203 16:57:11.239828 4998 scope.go:117] "RemoveContainer" containerID="374464700e2b9db2be22bbb601cba0881bf1cbf484fbf47de550ab712a2fdb6b" Dec 03 16:57:11 crc kubenswrapper[4998]: E1203 16:57:11.241314 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"374464700e2b9db2be22bbb601cba0881bf1cbf484fbf47de550ab712a2fdb6b\": container with ID starting with 374464700e2b9db2be22bbb601cba0881bf1cbf484fbf47de550ab712a2fdb6b not found: ID does not exist" containerID="374464700e2b9db2be22bbb601cba0881bf1cbf484fbf47de550ab712a2fdb6b" Dec 03 16:57:11 crc kubenswrapper[4998]: I1203 16:57:11.241384 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"374464700e2b9db2be22bbb601cba0881bf1cbf484fbf47de550ab712a2fdb6b"} err="failed to get container status \"374464700e2b9db2be22bbb601cba0881bf1cbf484fbf47de550ab712a2fdb6b\": rpc error: code = NotFound desc = could not find container \"374464700e2b9db2be22bbb601cba0881bf1cbf484fbf47de550ab712a2fdb6b\": container with ID starting with 374464700e2b9db2be22bbb601cba0881bf1cbf484fbf47de550ab712a2fdb6b not found: ID does not exist" Dec 03 16:57:11 crc kubenswrapper[4998]: I1203 16:57:11.691670 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a6110f-3034-45bd-89cc-176b1a7a2cb7" path="/var/lib/kubelet/pods/57a6110f-3034-45bd-89cc-176b1a7a2cb7/volumes" Dec 03 16:58:01 crc kubenswrapper[4998]: I1203 16:58:01.314442 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-frz5w"] Dec 03 16:58:01 crc kubenswrapper[4998]: E1203 16:58:01.315472 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57a6110f-3034-45bd-89cc-176b1a7a2cb7" containerName="extract-utilities" Dec 03 16:58:01 crc kubenswrapper[4998]: I1203 16:58:01.315488 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="57a6110f-3034-45bd-89cc-176b1a7a2cb7" containerName="extract-utilities" Dec 03 16:58:01 crc kubenswrapper[4998]: E1203 16:58:01.315520 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57a6110f-3034-45bd-89cc-176b1a7a2cb7" containerName="registry-server" Dec 03 16:58:01 crc kubenswrapper[4998]: I1203 16:58:01.315529 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="57a6110f-3034-45bd-89cc-176b1a7a2cb7" containerName="registry-server" Dec 03 16:58:01 crc kubenswrapper[4998]: E1203 16:58:01.315556 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57a6110f-3034-45bd-89cc-176b1a7a2cb7" containerName="extract-content" Dec 03 16:58:01 crc kubenswrapper[4998]: I1203 16:58:01.315565 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="57a6110f-3034-45bd-89cc-176b1a7a2cb7" containerName="extract-content" Dec 03 16:58:01 crc kubenswrapper[4998]: I1203 16:58:01.318394 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="57a6110f-3034-45bd-89cc-176b1a7a2cb7" containerName="registry-server" Dec 03 16:58:01 crc kubenswrapper[4998]: I1203 16:58:01.320903 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frz5w" Dec 03 16:58:01 crc kubenswrapper[4998]: I1203 16:58:01.335797 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-frz5w"] Dec 03 16:58:01 crc kubenswrapper[4998]: I1203 16:58:01.401472 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b49f9b1-7d88-4159-8736-336cf89f1623-catalog-content\") pod \"redhat-operators-frz5w\" (UID: \"4b49f9b1-7d88-4159-8736-336cf89f1623\") " pod="openshift-marketplace/redhat-operators-frz5w" Dec 03 16:58:01 crc kubenswrapper[4998]: I1203 16:58:01.401544 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b49f9b1-7d88-4159-8736-336cf89f1623-utilities\") pod \"redhat-operators-frz5w\" (UID: \"4b49f9b1-7d88-4159-8736-336cf89f1623\") " pod="openshift-marketplace/redhat-operators-frz5w" Dec 03 16:58:01 crc kubenswrapper[4998]: I1203 16:58:01.401654 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvgsk\" (UniqueName: \"kubernetes.io/projected/4b49f9b1-7d88-4159-8736-336cf89f1623-kube-api-access-wvgsk\") pod \"redhat-operators-frz5w\" (UID: \"4b49f9b1-7d88-4159-8736-336cf89f1623\") " pod="openshift-marketplace/redhat-operators-frz5w" Dec 03 16:58:01 crc kubenswrapper[4998]: I1203 16:58:01.503773 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b49f9b1-7d88-4159-8736-336cf89f1623-catalog-content\") pod \"redhat-operators-frz5w\" (UID: \"4b49f9b1-7d88-4159-8736-336cf89f1623\") " pod="openshift-marketplace/redhat-operators-frz5w" Dec 03 16:58:01 crc kubenswrapper[4998]: I1203 16:58:01.503836 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b49f9b1-7d88-4159-8736-336cf89f1623-utilities\") pod \"redhat-operators-frz5w\" (UID: \"4b49f9b1-7d88-4159-8736-336cf89f1623\") " pod="openshift-marketplace/redhat-operators-frz5w" Dec 03 16:58:01 crc kubenswrapper[4998]: I1203 16:58:01.503975 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvgsk\" (UniqueName: \"kubernetes.io/projected/4b49f9b1-7d88-4159-8736-336cf89f1623-kube-api-access-wvgsk\") pod \"redhat-operators-frz5w\" (UID: \"4b49f9b1-7d88-4159-8736-336cf89f1623\") " pod="openshift-marketplace/redhat-operators-frz5w" Dec 03 16:58:01 crc kubenswrapper[4998]: I1203 16:58:01.504493 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b49f9b1-7d88-4159-8736-336cf89f1623-utilities\") pod \"redhat-operators-frz5w\" (UID: \"4b49f9b1-7d88-4159-8736-336cf89f1623\") " pod="openshift-marketplace/redhat-operators-frz5w" Dec 03 16:58:01 crc kubenswrapper[4998]: I1203 16:58:01.504504 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b49f9b1-7d88-4159-8736-336cf89f1623-catalog-content\") pod \"redhat-operators-frz5w\" (UID: \"4b49f9b1-7d88-4159-8736-336cf89f1623\") " pod="openshift-marketplace/redhat-operators-frz5w" Dec 03 16:58:01 crc kubenswrapper[4998]: I1203 16:58:01.523800 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvgsk\" (UniqueName: \"kubernetes.io/projected/4b49f9b1-7d88-4159-8736-336cf89f1623-kube-api-access-wvgsk\") pod \"redhat-operators-frz5w\" (UID: \"4b49f9b1-7d88-4159-8736-336cf89f1623\") " pod="openshift-marketplace/redhat-operators-frz5w" Dec 03 16:58:01 crc kubenswrapper[4998]: I1203 16:58:01.672194 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frz5w" Dec 03 16:58:02 crc kubenswrapper[4998]: I1203 16:58:02.249013 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-frz5w"] Dec 03 16:58:03 crc kubenswrapper[4998]: I1203 16:58:03.068565 4998 generic.go:334] "Generic (PLEG): container finished" podID="4b49f9b1-7d88-4159-8736-336cf89f1623" containerID="4d8d3a9f48cd207cc6945eeec26a91d258a082ef630145e0b465fb5bb4139e09" exitCode=0 Dec 03 16:58:03 crc kubenswrapper[4998]: I1203 16:58:03.068684 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frz5w" event={"ID":"4b49f9b1-7d88-4159-8736-336cf89f1623","Type":"ContainerDied","Data":"4d8d3a9f48cd207cc6945eeec26a91d258a082ef630145e0b465fb5bb4139e09"} Dec 03 16:58:03 crc kubenswrapper[4998]: I1203 16:58:03.068954 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frz5w" event={"ID":"4b49f9b1-7d88-4159-8736-336cf89f1623","Type":"ContainerStarted","Data":"6bc21d7ff93633a022f2277a261b40466aca1f21db77ba30d8991e4a8bbc87f1"} Dec 03 16:58:05 crc kubenswrapper[4998]: I1203 16:58:05.091319 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frz5w" event={"ID":"4b49f9b1-7d88-4159-8736-336cf89f1623","Type":"ContainerStarted","Data":"9825e1acd6f355c5adb95daf0f7b2f9b514f3e86f2197abc275fb4fd0abaebfc"} Dec 03 16:58:09 crc kubenswrapper[4998]: I1203 16:58:09.133209 4998 generic.go:334] "Generic (PLEG): container finished" podID="4b49f9b1-7d88-4159-8736-336cf89f1623" containerID="9825e1acd6f355c5adb95daf0f7b2f9b514f3e86f2197abc275fb4fd0abaebfc" exitCode=0 Dec 03 16:58:09 crc kubenswrapper[4998]: I1203 16:58:09.133305 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frz5w" event={"ID":"4b49f9b1-7d88-4159-8736-336cf89f1623","Type":"ContainerDied","Data":"9825e1acd6f355c5adb95daf0f7b2f9b514f3e86f2197abc275fb4fd0abaebfc"} Dec 03 16:58:10 crc kubenswrapper[4998]: I1203 16:58:10.145296 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frz5w" event={"ID":"4b49f9b1-7d88-4159-8736-336cf89f1623","Type":"ContainerStarted","Data":"c9130537f841695f26ebd6c10c8973e04ff42a037bad5ca85685d33a98b375e3"} Dec 03 16:58:10 crc kubenswrapper[4998]: I1203 16:58:10.197369 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-frz5w" podStartSLOduration=2.5403104450000003 podStartE2EDuration="9.197343798s" podCreationTimestamp="2025-12-03 16:58:01 +0000 UTC" firstStartedPulling="2025-12-03 16:58:03.070609132 +0000 UTC m=+3261.682309365" lastFinishedPulling="2025-12-03 16:58:09.727642455 +0000 UTC m=+3268.339342718" observedRunningTime="2025-12-03 16:58:10.160441113 +0000 UTC m=+3268.772141336" watchObservedRunningTime="2025-12-03 16:58:10.197343798 +0000 UTC m=+3268.809044021" Dec 03 16:58:11 crc kubenswrapper[4998]: I1203 16:58:11.673889 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-frz5w" Dec 03 16:58:11 crc kubenswrapper[4998]: I1203 16:58:11.674668 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-frz5w" Dec 03 16:58:12 crc kubenswrapper[4998]: I1203 16:58:12.726505 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-frz5w" podUID="4b49f9b1-7d88-4159-8736-336cf89f1623" containerName="registry-server" probeResult="failure" output=< Dec 03 16:58:12 crc kubenswrapper[4998]: timeout: failed to connect service ":50051" within 1s Dec 03 16:58:12 crc kubenswrapper[4998]: > Dec 03 16:58:21 crc kubenswrapper[4998]: I1203 16:58:21.763081 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-frz5w" Dec 03 16:58:21 crc kubenswrapper[4998]: I1203 16:58:21.842363 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-frz5w" Dec 03 16:58:22 crc kubenswrapper[4998]: I1203 16:58:22.002291 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-frz5w"] Dec 03 16:58:23 crc kubenswrapper[4998]: I1203 16:58:23.307175 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-frz5w" podUID="4b49f9b1-7d88-4159-8736-336cf89f1623" containerName="registry-server" containerID="cri-o://c9130537f841695f26ebd6c10c8973e04ff42a037bad5ca85685d33a98b375e3" gracePeriod=2 Dec 03 16:58:23 crc kubenswrapper[4998]: I1203 16:58:23.972517 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frz5w" Dec 03 16:58:24 crc kubenswrapper[4998]: I1203 16:58:24.008669 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvgsk\" (UniqueName: \"kubernetes.io/projected/4b49f9b1-7d88-4159-8736-336cf89f1623-kube-api-access-wvgsk\") pod \"4b49f9b1-7d88-4159-8736-336cf89f1623\" (UID: \"4b49f9b1-7d88-4159-8736-336cf89f1623\") " Dec 03 16:58:24 crc kubenswrapper[4998]: I1203 16:58:24.008837 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b49f9b1-7d88-4159-8736-336cf89f1623-catalog-content\") pod \"4b49f9b1-7d88-4159-8736-336cf89f1623\" (UID: \"4b49f9b1-7d88-4159-8736-336cf89f1623\") " Dec 03 16:58:24 crc kubenswrapper[4998]: I1203 16:58:24.009001 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b49f9b1-7d88-4159-8736-336cf89f1623-utilities\") pod \"4b49f9b1-7d88-4159-8736-336cf89f1623\" (UID: \"4b49f9b1-7d88-4159-8736-336cf89f1623\") " Dec 03 16:58:24 crc kubenswrapper[4998]: I1203 16:58:24.010528 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b49f9b1-7d88-4159-8736-336cf89f1623-utilities" (OuterVolumeSpecName: "utilities") pod "4b49f9b1-7d88-4159-8736-336cf89f1623" (UID: "4b49f9b1-7d88-4159-8736-336cf89f1623"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:58:24 crc kubenswrapper[4998]: I1203 16:58:24.025972 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b49f9b1-7d88-4159-8736-336cf89f1623-kube-api-access-wvgsk" (OuterVolumeSpecName: "kube-api-access-wvgsk") pod "4b49f9b1-7d88-4159-8736-336cf89f1623" (UID: "4b49f9b1-7d88-4159-8736-336cf89f1623"). InnerVolumeSpecName "kube-api-access-wvgsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 16:58:24 crc kubenswrapper[4998]: I1203 16:58:24.111246 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b49f9b1-7d88-4159-8736-336cf89f1623-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 16:58:24 crc kubenswrapper[4998]: I1203 16:58:24.111480 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvgsk\" (UniqueName: \"kubernetes.io/projected/4b49f9b1-7d88-4159-8736-336cf89f1623-kube-api-access-wvgsk\") on node \"crc\" DevicePath \"\"" Dec 03 16:58:24 crc kubenswrapper[4998]: I1203 16:58:24.129927 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b49f9b1-7d88-4159-8736-336cf89f1623-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4b49f9b1-7d88-4159-8736-336cf89f1623" (UID: "4b49f9b1-7d88-4159-8736-336cf89f1623"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 16:58:24 crc kubenswrapper[4998]: I1203 16:58:24.213314 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b49f9b1-7d88-4159-8736-336cf89f1623-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 16:58:24 crc kubenswrapper[4998]: I1203 16:58:24.317081 4998 generic.go:334] "Generic (PLEG): container finished" podID="4b49f9b1-7d88-4159-8736-336cf89f1623" containerID="c9130537f841695f26ebd6c10c8973e04ff42a037bad5ca85685d33a98b375e3" exitCode=0 Dec 03 16:58:24 crc kubenswrapper[4998]: I1203 16:58:24.317127 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frz5w" event={"ID":"4b49f9b1-7d88-4159-8736-336cf89f1623","Type":"ContainerDied","Data":"c9130537f841695f26ebd6c10c8973e04ff42a037bad5ca85685d33a98b375e3"} Dec 03 16:58:24 crc kubenswrapper[4998]: I1203 16:58:24.317151 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-frz5w" event={"ID":"4b49f9b1-7d88-4159-8736-336cf89f1623","Type":"ContainerDied","Data":"6bc21d7ff93633a022f2277a261b40466aca1f21db77ba30d8991e4a8bbc87f1"} Dec 03 16:58:24 crc kubenswrapper[4998]: I1203 16:58:24.317169 4998 scope.go:117] "RemoveContainer" containerID="c9130537f841695f26ebd6c10c8973e04ff42a037bad5ca85685d33a98b375e3" Dec 03 16:58:24 crc kubenswrapper[4998]: I1203 16:58:24.317269 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-frz5w" Dec 03 16:58:24 crc kubenswrapper[4998]: I1203 16:58:24.346850 4998 scope.go:117] "RemoveContainer" containerID="9825e1acd6f355c5adb95daf0f7b2f9b514f3e86f2197abc275fb4fd0abaebfc" Dec 03 16:58:24 crc kubenswrapper[4998]: I1203 16:58:24.349892 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-frz5w"] Dec 03 16:58:24 crc kubenswrapper[4998]: I1203 16:58:24.359598 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-frz5w"] Dec 03 16:58:24 crc kubenswrapper[4998]: I1203 16:58:24.403035 4998 scope.go:117] "RemoveContainer" containerID="4d8d3a9f48cd207cc6945eeec26a91d258a082ef630145e0b465fb5bb4139e09" Dec 03 16:58:24 crc kubenswrapper[4998]: I1203 16:58:24.480477 4998 scope.go:117] "RemoveContainer" containerID="c9130537f841695f26ebd6c10c8973e04ff42a037bad5ca85685d33a98b375e3" Dec 03 16:58:24 crc kubenswrapper[4998]: E1203 16:58:24.481045 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9130537f841695f26ebd6c10c8973e04ff42a037bad5ca85685d33a98b375e3\": container with ID starting with c9130537f841695f26ebd6c10c8973e04ff42a037bad5ca85685d33a98b375e3 not found: ID does not exist" containerID="c9130537f841695f26ebd6c10c8973e04ff42a037bad5ca85685d33a98b375e3" Dec 03 16:58:24 crc kubenswrapper[4998]: I1203 16:58:24.481086 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9130537f841695f26ebd6c10c8973e04ff42a037bad5ca85685d33a98b375e3"} err="failed to get container status \"c9130537f841695f26ebd6c10c8973e04ff42a037bad5ca85685d33a98b375e3\": rpc error: code = NotFound desc = could not find container \"c9130537f841695f26ebd6c10c8973e04ff42a037bad5ca85685d33a98b375e3\": container with ID starting with c9130537f841695f26ebd6c10c8973e04ff42a037bad5ca85685d33a98b375e3 not found: ID does not exist" Dec 03 16:58:24 crc kubenswrapper[4998]: I1203 16:58:24.481114 4998 scope.go:117] "RemoveContainer" containerID="9825e1acd6f355c5adb95daf0f7b2f9b514f3e86f2197abc275fb4fd0abaebfc" Dec 03 16:58:24 crc kubenswrapper[4998]: E1203 16:58:24.481416 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9825e1acd6f355c5adb95daf0f7b2f9b514f3e86f2197abc275fb4fd0abaebfc\": container with ID starting with 9825e1acd6f355c5adb95daf0f7b2f9b514f3e86f2197abc275fb4fd0abaebfc not found: ID does not exist" containerID="9825e1acd6f355c5adb95daf0f7b2f9b514f3e86f2197abc275fb4fd0abaebfc" Dec 03 16:58:24 crc kubenswrapper[4998]: I1203 16:58:24.481445 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9825e1acd6f355c5adb95daf0f7b2f9b514f3e86f2197abc275fb4fd0abaebfc"} err="failed to get container status \"9825e1acd6f355c5adb95daf0f7b2f9b514f3e86f2197abc275fb4fd0abaebfc\": rpc error: code = NotFound desc = could not find container \"9825e1acd6f355c5adb95daf0f7b2f9b514f3e86f2197abc275fb4fd0abaebfc\": container with ID starting with 9825e1acd6f355c5adb95daf0f7b2f9b514f3e86f2197abc275fb4fd0abaebfc not found: ID does not exist" Dec 03 16:58:24 crc kubenswrapper[4998]: I1203 16:58:24.481467 4998 scope.go:117] "RemoveContainer" containerID="4d8d3a9f48cd207cc6945eeec26a91d258a082ef630145e0b465fb5bb4139e09" Dec 03 16:58:24 crc kubenswrapper[4998]: E1203 16:58:24.481664 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d8d3a9f48cd207cc6945eeec26a91d258a082ef630145e0b465fb5bb4139e09\": container with ID starting with 4d8d3a9f48cd207cc6945eeec26a91d258a082ef630145e0b465fb5bb4139e09 not found: ID does not exist" containerID="4d8d3a9f48cd207cc6945eeec26a91d258a082ef630145e0b465fb5bb4139e09" Dec 03 16:58:24 crc kubenswrapper[4998]: I1203 16:58:24.481683 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d8d3a9f48cd207cc6945eeec26a91d258a082ef630145e0b465fb5bb4139e09"} err="failed to get container status \"4d8d3a9f48cd207cc6945eeec26a91d258a082ef630145e0b465fb5bb4139e09\": rpc error: code = NotFound desc = could not find container \"4d8d3a9f48cd207cc6945eeec26a91d258a082ef630145e0b465fb5bb4139e09\": container with ID starting with 4d8d3a9f48cd207cc6945eeec26a91d258a082ef630145e0b465fb5bb4139e09 not found: ID does not exist" Dec 03 16:58:25 crc kubenswrapper[4998]: I1203 16:58:25.694228 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b49f9b1-7d88-4159-8736-336cf89f1623" path="/var/lib/kubelet/pods/4b49f9b1-7d88-4159-8736-336cf89f1623/volumes" Dec 03 16:58:57 crc kubenswrapper[4998]: I1203 16:58:57.111332 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:58:57 crc kubenswrapper[4998]: I1203 16:58:57.112105 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:59:27 crc kubenswrapper[4998]: I1203 16:59:27.112099 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:59:27 crc kubenswrapper[4998]: I1203 16:59:27.112957 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:59:57 crc kubenswrapper[4998]: I1203 16:59:57.110816 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 16:59:57 crc kubenswrapper[4998]: I1203 16:59:57.111413 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 16:59:57 crc kubenswrapper[4998]: I1203 16:59:57.111465 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 16:59:57 crc kubenswrapper[4998]: I1203 16:59:57.112287 4998 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"84aefa566144217de53fa4d3aefaa1bc2dcd21fcd378c021e19510c75d4d205e"} pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 16:59:57 crc kubenswrapper[4998]: I1203 16:59:57.112339 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" containerID="cri-o://84aefa566144217de53fa4d3aefaa1bc2dcd21fcd378c021e19510c75d4d205e" gracePeriod=600 Dec 03 16:59:57 crc kubenswrapper[4998]: I1203 16:59:57.326706 4998 generic.go:334] "Generic (PLEG): container finished" podID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerID="84aefa566144217de53fa4d3aefaa1bc2dcd21fcd378c021e19510c75d4d205e" exitCode=0 Dec 03 16:59:57 crc kubenswrapper[4998]: I1203 16:59:57.326806 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerDied","Data":"84aefa566144217de53fa4d3aefaa1bc2dcd21fcd378c021e19510c75d4d205e"} Dec 03 16:59:57 crc kubenswrapper[4998]: I1203 16:59:57.327050 4998 scope.go:117] "RemoveContainer" containerID="21aa41d7e16b4544c102c8fff03f831c7d60d95e61bff2aff971a3425d442096" Dec 03 16:59:58 crc kubenswrapper[4998]: I1203 16:59:58.337376 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14"} Dec 03 17:00:00 crc kubenswrapper[4998]: I1203 17:00:00.142691 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413020-j87g2"] Dec 03 17:00:00 crc kubenswrapper[4998]: E1203 17:00:00.143732 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b49f9b1-7d88-4159-8736-336cf89f1623" containerName="extract-utilities" Dec 03 17:00:00 crc kubenswrapper[4998]: I1203 17:00:00.143765 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b49f9b1-7d88-4159-8736-336cf89f1623" containerName="extract-utilities" Dec 03 17:00:00 crc kubenswrapper[4998]: E1203 17:00:00.143796 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b49f9b1-7d88-4159-8736-336cf89f1623" containerName="extract-content" Dec 03 17:00:00 crc kubenswrapper[4998]: I1203 17:00:00.143804 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b49f9b1-7d88-4159-8736-336cf89f1623" containerName="extract-content" Dec 03 17:00:00 crc kubenswrapper[4998]: E1203 17:00:00.143823 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b49f9b1-7d88-4159-8736-336cf89f1623" containerName="registry-server" Dec 03 17:00:00 crc kubenswrapper[4998]: I1203 17:00:00.143833 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b49f9b1-7d88-4159-8736-336cf89f1623" containerName="registry-server" Dec 03 17:00:00 crc kubenswrapper[4998]: I1203 17:00:00.144098 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b49f9b1-7d88-4159-8736-336cf89f1623" containerName="registry-server" Dec 03 17:00:00 crc kubenswrapper[4998]: I1203 17:00:00.145011 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-j87g2" Dec 03 17:00:00 crc kubenswrapper[4998]: I1203 17:00:00.148009 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 17:00:00 crc kubenswrapper[4998]: I1203 17:00:00.153976 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413020-j87g2"] Dec 03 17:00:00 crc kubenswrapper[4998]: I1203 17:00:00.164949 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 17:00:00 crc kubenswrapper[4998]: I1203 17:00:00.251981 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/026f9ce6-dbab-4142-8c66-3035f53b8054-config-volume\") pod \"collect-profiles-29413020-j87g2\" (UID: \"026f9ce6-dbab-4142-8c66-3035f53b8054\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-j87g2" Dec 03 17:00:00 crc kubenswrapper[4998]: I1203 17:00:00.252102 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/026f9ce6-dbab-4142-8c66-3035f53b8054-secret-volume\") pod \"collect-profiles-29413020-j87g2\" (UID: \"026f9ce6-dbab-4142-8c66-3035f53b8054\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-j87g2" Dec 03 17:00:00 crc kubenswrapper[4998]: I1203 17:00:00.252141 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrzbw\" (UniqueName: \"kubernetes.io/projected/026f9ce6-dbab-4142-8c66-3035f53b8054-kube-api-access-hrzbw\") pod \"collect-profiles-29413020-j87g2\" (UID: \"026f9ce6-dbab-4142-8c66-3035f53b8054\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-j87g2" Dec 03 17:00:00 crc kubenswrapper[4998]: I1203 17:00:00.354618 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/026f9ce6-dbab-4142-8c66-3035f53b8054-secret-volume\") pod \"collect-profiles-29413020-j87g2\" (UID: \"026f9ce6-dbab-4142-8c66-3035f53b8054\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-j87g2" Dec 03 17:00:00 crc kubenswrapper[4998]: I1203 17:00:00.355667 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrzbw\" (UniqueName: \"kubernetes.io/projected/026f9ce6-dbab-4142-8c66-3035f53b8054-kube-api-access-hrzbw\") pod \"collect-profiles-29413020-j87g2\" (UID: \"026f9ce6-dbab-4142-8c66-3035f53b8054\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-j87g2" Dec 03 17:00:00 crc kubenswrapper[4998]: I1203 17:00:00.355918 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/026f9ce6-dbab-4142-8c66-3035f53b8054-config-volume\") pod \"collect-profiles-29413020-j87g2\" (UID: \"026f9ce6-dbab-4142-8c66-3035f53b8054\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-j87g2" Dec 03 17:00:00 crc kubenswrapper[4998]: I1203 17:00:00.356623 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/026f9ce6-dbab-4142-8c66-3035f53b8054-config-volume\") pod \"collect-profiles-29413020-j87g2\" (UID: \"026f9ce6-dbab-4142-8c66-3035f53b8054\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-j87g2" Dec 03 17:00:00 crc kubenswrapper[4998]: I1203 17:00:00.361513 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/026f9ce6-dbab-4142-8c66-3035f53b8054-secret-volume\") pod \"collect-profiles-29413020-j87g2\" (UID: \"026f9ce6-dbab-4142-8c66-3035f53b8054\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-j87g2" Dec 03 17:00:00 crc kubenswrapper[4998]: I1203 17:00:00.374723 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrzbw\" (UniqueName: \"kubernetes.io/projected/026f9ce6-dbab-4142-8c66-3035f53b8054-kube-api-access-hrzbw\") pod \"collect-profiles-29413020-j87g2\" (UID: \"026f9ce6-dbab-4142-8c66-3035f53b8054\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-j87g2" Dec 03 17:00:00 crc kubenswrapper[4998]: I1203 17:00:00.491628 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-j87g2" Dec 03 17:00:00 crc kubenswrapper[4998]: I1203 17:00:00.944845 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413020-j87g2"] Dec 03 17:00:01 crc kubenswrapper[4998]: I1203 17:00:01.372377 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-j87g2" event={"ID":"026f9ce6-dbab-4142-8c66-3035f53b8054","Type":"ContainerStarted","Data":"80e43abb86da11c9f5c02522e86ce07e1497f34e8a79d05fbd7e93e9c2997bbf"} Dec 03 17:00:01 crc kubenswrapper[4998]: I1203 17:00:01.373802 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-j87g2" event={"ID":"026f9ce6-dbab-4142-8c66-3035f53b8054","Type":"ContainerStarted","Data":"0854bb25d0017bf4da9532347c4e79d399abe55233c4c1e518d2496fe730cb28"} Dec 03 17:00:02 crc kubenswrapper[4998]: I1203 17:00:02.382703 4998 generic.go:334] "Generic (PLEG): container finished" podID="026f9ce6-dbab-4142-8c66-3035f53b8054" containerID="80e43abb86da11c9f5c02522e86ce07e1497f34e8a79d05fbd7e93e9c2997bbf" exitCode=0 Dec 03 17:00:02 crc kubenswrapper[4998]: I1203 17:00:02.382812 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-j87g2" event={"ID":"026f9ce6-dbab-4142-8c66-3035f53b8054","Type":"ContainerDied","Data":"80e43abb86da11c9f5c02522e86ce07e1497f34e8a79d05fbd7e93e9c2997bbf"} Dec 03 17:00:03 crc kubenswrapper[4998]: I1203 17:00:03.864123 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-j87g2" Dec 03 17:00:04 crc kubenswrapper[4998]: I1203 17:00:04.050634 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrzbw\" (UniqueName: \"kubernetes.io/projected/026f9ce6-dbab-4142-8c66-3035f53b8054-kube-api-access-hrzbw\") pod \"026f9ce6-dbab-4142-8c66-3035f53b8054\" (UID: \"026f9ce6-dbab-4142-8c66-3035f53b8054\") " Dec 03 17:00:04 crc kubenswrapper[4998]: I1203 17:00:04.051018 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/026f9ce6-dbab-4142-8c66-3035f53b8054-config-volume\") pod \"026f9ce6-dbab-4142-8c66-3035f53b8054\" (UID: \"026f9ce6-dbab-4142-8c66-3035f53b8054\") " Dec 03 17:00:04 crc kubenswrapper[4998]: I1203 17:00:04.051161 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/026f9ce6-dbab-4142-8c66-3035f53b8054-secret-volume\") pod \"026f9ce6-dbab-4142-8c66-3035f53b8054\" (UID: \"026f9ce6-dbab-4142-8c66-3035f53b8054\") " Dec 03 17:00:04 crc kubenswrapper[4998]: I1203 17:00:04.052911 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/026f9ce6-dbab-4142-8c66-3035f53b8054-config-volume" (OuterVolumeSpecName: "config-volume") pod "026f9ce6-dbab-4142-8c66-3035f53b8054" (UID: "026f9ce6-dbab-4142-8c66-3035f53b8054"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:00:04 crc kubenswrapper[4998]: I1203 17:00:04.058135 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/026f9ce6-dbab-4142-8c66-3035f53b8054-kube-api-access-hrzbw" (OuterVolumeSpecName: "kube-api-access-hrzbw") pod "026f9ce6-dbab-4142-8c66-3035f53b8054" (UID: "026f9ce6-dbab-4142-8c66-3035f53b8054"). InnerVolumeSpecName "kube-api-access-hrzbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:00:04 crc kubenswrapper[4998]: I1203 17:00:04.061384 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/026f9ce6-dbab-4142-8c66-3035f53b8054-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "026f9ce6-dbab-4142-8c66-3035f53b8054" (UID: "026f9ce6-dbab-4142-8c66-3035f53b8054"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:00:04 crc kubenswrapper[4998]: I1203 17:00:04.153219 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrzbw\" (UniqueName: \"kubernetes.io/projected/026f9ce6-dbab-4142-8c66-3035f53b8054-kube-api-access-hrzbw\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:04 crc kubenswrapper[4998]: I1203 17:00:04.153264 4998 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/026f9ce6-dbab-4142-8c66-3035f53b8054-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:04 crc kubenswrapper[4998]: I1203 17:00:04.153276 4998 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/026f9ce6-dbab-4142-8c66-3035f53b8054-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:04 crc kubenswrapper[4998]: I1203 17:00:04.403966 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-j87g2" event={"ID":"026f9ce6-dbab-4142-8c66-3035f53b8054","Type":"ContainerDied","Data":"0854bb25d0017bf4da9532347c4e79d399abe55233c4c1e518d2496fe730cb28"} Dec 03 17:00:04 crc kubenswrapper[4998]: I1203 17:00:04.404014 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0854bb25d0017bf4da9532347c4e79d399abe55233c4c1e518d2496fe730cb28" Dec 03 17:00:04 crc kubenswrapper[4998]: I1203 17:00:04.404048 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413020-j87g2" Dec 03 17:00:04 crc kubenswrapper[4998]: I1203 17:00:04.952609 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412975-hfbt9"] Dec 03 17:00:04 crc kubenswrapper[4998]: I1203 17:00:04.961424 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412975-hfbt9"] Dec 03 17:00:05 crc kubenswrapper[4998]: I1203 17:00:05.689349 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f249b9c6-1ad9-44b0-8475-8c11dd212f01" path="/var/lib/kubelet/pods/f249b9c6-1ad9-44b0-8475-8c11dd212f01/volumes" Dec 03 17:00:07 crc kubenswrapper[4998]: I1203 17:00:07.314482 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t6bk2"] Dec 03 17:00:07 crc kubenswrapper[4998]: E1203 17:00:07.315500 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="026f9ce6-dbab-4142-8c66-3035f53b8054" containerName="collect-profiles" Dec 03 17:00:07 crc kubenswrapper[4998]: I1203 17:00:07.315521 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="026f9ce6-dbab-4142-8c66-3035f53b8054" containerName="collect-profiles" Dec 03 17:00:07 crc kubenswrapper[4998]: I1203 17:00:07.315912 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="026f9ce6-dbab-4142-8c66-3035f53b8054" containerName="collect-profiles" Dec 03 17:00:07 crc kubenswrapper[4998]: I1203 17:00:07.318967 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t6bk2" Dec 03 17:00:07 crc kubenswrapper[4998]: I1203 17:00:07.325934 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t6bk2"] Dec 03 17:00:07 crc kubenswrapper[4998]: I1203 17:00:07.332264 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/999ac090-324b-4348-8feb-215ecb6e509b-catalog-content\") pod \"community-operators-t6bk2\" (UID: \"999ac090-324b-4348-8feb-215ecb6e509b\") " pod="openshift-marketplace/community-operators-t6bk2" Dec 03 17:00:07 crc kubenswrapper[4998]: I1203 17:00:07.332576 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fnwj\" (UniqueName: \"kubernetes.io/projected/999ac090-324b-4348-8feb-215ecb6e509b-kube-api-access-6fnwj\") pod \"community-operators-t6bk2\" (UID: \"999ac090-324b-4348-8feb-215ecb6e509b\") " pod="openshift-marketplace/community-operators-t6bk2" Dec 03 17:00:07 crc kubenswrapper[4998]: I1203 17:00:07.332865 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/999ac090-324b-4348-8feb-215ecb6e509b-utilities\") pod \"community-operators-t6bk2\" (UID: \"999ac090-324b-4348-8feb-215ecb6e509b\") " pod="openshift-marketplace/community-operators-t6bk2" Dec 03 17:00:07 crc kubenswrapper[4998]: I1203 17:00:07.435383 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/999ac090-324b-4348-8feb-215ecb6e509b-catalog-content\") pod \"community-operators-t6bk2\" (UID: \"999ac090-324b-4348-8feb-215ecb6e509b\") " pod="openshift-marketplace/community-operators-t6bk2" Dec 03 17:00:07 crc kubenswrapper[4998]: I1203 17:00:07.435526 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fnwj\" (UniqueName: \"kubernetes.io/projected/999ac090-324b-4348-8feb-215ecb6e509b-kube-api-access-6fnwj\") pod \"community-operators-t6bk2\" (UID: \"999ac090-324b-4348-8feb-215ecb6e509b\") " pod="openshift-marketplace/community-operators-t6bk2" Dec 03 17:00:07 crc kubenswrapper[4998]: I1203 17:00:07.435577 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/999ac090-324b-4348-8feb-215ecb6e509b-utilities\") pod \"community-operators-t6bk2\" (UID: \"999ac090-324b-4348-8feb-215ecb6e509b\") " pod="openshift-marketplace/community-operators-t6bk2" Dec 03 17:00:07 crc kubenswrapper[4998]: I1203 17:00:07.436040 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/999ac090-324b-4348-8feb-215ecb6e509b-catalog-content\") pod \"community-operators-t6bk2\" (UID: \"999ac090-324b-4348-8feb-215ecb6e509b\") " pod="openshift-marketplace/community-operators-t6bk2" Dec 03 17:00:07 crc kubenswrapper[4998]: I1203 17:00:07.436070 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/999ac090-324b-4348-8feb-215ecb6e509b-utilities\") pod \"community-operators-t6bk2\" (UID: \"999ac090-324b-4348-8feb-215ecb6e509b\") " pod="openshift-marketplace/community-operators-t6bk2" Dec 03 17:00:07 crc kubenswrapper[4998]: I1203 17:00:07.456816 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fnwj\" (UniqueName: \"kubernetes.io/projected/999ac090-324b-4348-8feb-215ecb6e509b-kube-api-access-6fnwj\") pod \"community-operators-t6bk2\" (UID: \"999ac090-324b-4348-8feb-215ecb6e509b\") " pod="openshift-marketplace/community-operators-t6bk2" Dec 03 17:00:07 crc kubenswrapper[4998]: I1203 17:00:07.648979 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t6bk2" Dec 03 17:00:08 crc kubenswrapper[4998]: I1203 17:00:08.298302 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t6bk2"] Dec 03 17:00:08 crc kubenswrapper[4998]: I1203 17:00:08.440394 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t6bk2" event={"ID":"999ac090-324b-4348-8feb-215ecb6e509b","Type":"ContainerStarted","Data":"5c51ad8999cf64a360ffdabe65937cb6a17c6b3425f2e4ab3d4f1fb3baf0a7ce"} Dec 03 17:00:09 crc kubenswrapper[4998]: I1203 17:00:09.473365 4998 generic.go:334] "Generic (PLEG): container finished" podID="999ac090-324b-4348-8feb-215ecb6e509b" containerID="a39ed393243d9fd77c2c5a3bb547b07a8fa6eee34f409f61e397b9df111b5342" exitCode=0 Dec 03 17:00:09 crc kubenswrapper[4998]: I1203 17:00:09.473896 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t6bk2" event={"ID":"999ac090-324b-4348-8feb-215ecb6e509b","Type":"ContainerDied","Data":"a39ed393243d9fd77c2c5a3bb547b07a8fa6eee34f409f61e397b9df111b5342"} Dec 03 17:00:09 crc kubenswrapper[4998]: I1203 17:00:09.479199 4998 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 17:00:14 crc kubenswrapper[4998]: I1203 17:00:14.562739 4998 generic.go:334] "Generic (PLEG): container finished" podID="999ac090-324b-4348-8feb-215ecb6e509b" containerID="a0772bc0a7bda600ad4dfe295dac6f1c85991279717d5e3d8cb3ece4b52491a8" exitCode=0 Dec 03 17:00:14 crc kubenswrapper[4998]: I1203 17:00:14.562793 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t6bk2" event={"ID":"999ac090-324b-4348-8feb-215ecb6e509b","Type":"ContainerDied","Data":"a0772bc0a7bda600ad4dfe295dac6f1c85991279717d5e3d8cb3ece4b52491a8"} Dec 03 17:00:15 crc kubenswrapper[4998]: I1203 17:00:15.575220 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t6bk2" event={"ID":"999ac090-324b-4348-8feb-215ecb6e509b","Type":"ContainerStarted","Data":"6fa6227842ab4d807ead21843bef95d98fd45683ef9b7a142c5652848f501081"} Dec 03 17:00:15 crc kubenswrapper[4998]: I1203 17:00:15.603427 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t6bk2" podStartSLOduration=3.070290516 podStartE2EDuration="8.603405102s" podCreationTimestamp="2025-12-03 17:00:07 +0000 UTC" firstStartedPulling="2025-12-03 17:00:09.478924388 +0000 UTC m=+3388.090624611" lastFinishedPulling="2025-12-03 17:00:15.012038984 +0000 UTC m=+3393.623739197" observedRunningTime="2025-12-03 17:00:15.594292536 +0000 UTC m=+3394.205992809" watchObservedRunningTime="2025-12-03 17:00:15.603405102 +0000 UTC m=+3394.215105335" Dec 03 17:00:17 crc kubenswrapper[4998]: I1203 17:00:17.649317 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t6bk2" Dec 03 17:00:17 crc kubenswrapper[4998]: I1203 17:00:17.649902 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t6bk2" Dec 03 17:00:17 crc kubenswrapper[4998]: I1203 17:00:17.697499 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t6bk2" Dec 03 17:00:27 crc kubenswrapper[4998]: I1203 17:00:27.734693 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t6bk2" Dec 03 17:00:27 crc kubenswrapper[4998]: I1203 17:00:27.809209 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t6bk2"] Dec 03 17:00:27 crc kubenswrapper[4998]: I1203 17:00:27.860218 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tfv2f"] Dec 03 17:00:27 crc kubenswrapper[4998]: I1203 17:00:27.860497 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tfv2f" podUID="816953d5-a5ca-4822-94cb-600d1bbbdb6b" containerName="registry-server" containerID="cri-o://e63a0b66f917696b8b700cfb4f71a25bc43b3c43d5a39a4fbe5e36a7a33d867b" gracePeriod=2 Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.393839 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tfv2f" Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.446793 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgshq\" (UniqueName: \"kubernetes.io/projected/816953d5-a5ca-4822-94cb-600d1bbbdb6b-kube-api-access-wgshq\") pod \"816953d5-a5ca-4822-94cb-600d1bbbdb6b\" (UID: \"816953d5-a5ca-4822-94cb-600d1bbbdb6b\") " Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.446905 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/816953d5-a5ca-4822-94cb-600d1bbbdb6b-utilities\") pod \"816953d5-a5ca-4822-94cb-600d1bbbdb6b\" (UID: \"816953d5-a5ca-4822-94cb-600d1bbbdb6b\") " Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.446944 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/816953d5-a5ca-4822-94cb-600d1bbbdb6b-catalog-content\") pod \"816953d5-a5ca-4822-94cb-600d1bbbdb6b\" (UID: \"816953d5-a5ca-4822-94cb-600d1bbbdb6b\") " Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.449117 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/816953d5-a5ca-4822-94cb-600d1bbbdb6b-utilities" (OuterVolumeSpecName: "utilities") pod "816953d5-a5ca-4822-94cb-600d1bbbdb6b" (UID: "816953d5-a5ca-4822-94cb-600d1bbbdb6b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.465468 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/816953d5-a5ca-4822-94cb-600d1bbbdb6b-kube-api-access-wgshq" (OuterVolumeSpecName: "kube-api-access-wgshq") pod "816953d5-a5ca-4822-94cb-600d1bbbdb6b" (UID: "816953d5-a5ca-4822-94cb-600d1bbbdb6b"). InnerVolumeSpecName "kube-api-access-wgshq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.549111 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/816953d5-a5ca-4822-94cb-600d1bbbdb6b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "816953d5-a5ca-4822-94cb-600d1bbbdb6b" (UID: "816953d5-a5ca-4822-94cb-600d1bbbdb6b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.549648 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/816953d5-a5ca-4822-94cb-600d1bbbdb6b-catalog-content\") pod \"816953d5-a5ca-4822-94cb-600d1bbbdb6b\" (UID: \"816953d5-a5ca-4822-94cb-600d1bbbdb6b\") " Dec 03 17:00:28 crc kubenswrapper[4998]: W1203 17:00:28.549994 4998 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/816953d5-a5ca-4822-94cb-600d1bbbdb6b/volumes/kubernetes.io~empty-dir/catalog-content Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.550015 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/816953d5-a5ca-4822-94cb-600d1bbbdb6b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "816953d5-a5ca-4822-94cb-600d1bbbdb6b" (UID: "816953d5-a5ca-4822-94cb-600d1bbbdb6b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.550361 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgshq\" (UniqueName: \"kubernetes.io/projected/816953d5-a5ca-4822-94cb-600d1bbbdb6b-kube-api-access-wgshq\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.550384 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/816953d5-a5ca-4822-94cb-600d1bbbdb6b-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.550395 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/816953d5-a5ca-4822-94cb-600d1bbbdb6b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.745549 4998 generic.go:334] "Generic (PLEG): container finished" podID="816953d5-a5ca-4822-94cb-600d1bbbdb6b" containerID="e63a0b66f917696b8b700cfb4f71a25bc43b3c43d5a39a4fbe5e36a7a33d867b" exitCode=0 Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.745613 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tfv2f" Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.745631 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tfv2f" event={"ID":"816953d5-a5ca-4822-94cb-600d1bbbdb6b","Type":"ContainerDied","Data":"e63a0b66f917696b8b700cfb4f71a25bc43b3c43d5a39a4fbe5e36a7a33d867b"} Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.745674 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tfv2f" event={"ID":"816953d5-a5ca-4822-94cb-600d1bbbdb6b","Type":"ContainerDied","Data":"db2c1c6ea6bd617556048d515b79fa2af3379f7119fa1073ff736bf689f435de"} Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.745699 4998 scope.go:117] "RemoveContainer" containerID="e63a0b66f917696b8b700cfb4f71a25bc43b3c43d5a39a4fbe5e36a7a33d867b" Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.775011 4998 scope.go:117] "RemoveContainer" containerID="251c17f3e990734bc0e154eaaf214c1f5d15b8db902b3a9dc2903061c48c64c3" Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.801568 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tfv2f"] Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.835039 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tfv2f"] Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.839628 4998 scope.go:117] "RemoveContainer" containerID="adadcb4a6a35fa670d260da850a459953da744421077c24879c8a32cf37f0409" Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.883309 4998 scope.go:117] "RemoveContainer" containerID="e63a0b66f917696b8b700cfb4f71a25bc43b3c43d5a39a4fbe5e36a7a33d867b" Dec 03 17:00:28 crc kubenswrapper[4998]: E1203 17:00:28.883612 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e63a0b66f917696b8b700cfb4f71a25bc43b3c43d5a39a4fbe5e36a7a33d867b\": container with ID starting with e63a0b66f917696b8b700cfb4f71a25bc43b3c43d5a39a4fbe5e36a7a33d867b not found: ID does not exist" containerID="e63a0b66f917696b8b700cfb4f71a25bc43b3c43d5a39a4fbe5e36a7a33d867b" Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.883639 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e63a0b66f917696b8b700cfb4f71a25bc43b3c43d5a39a4fbe5e36a7a33d867b"} err="failed to get container status \"e63a0b66f917696b8b700cfb4f71a25bc43b3c43d5a39a4fbe5e36a7a33d867b\": rpc error: code = NotFound desc = could not find container \"e63a0b66f917696b8b700cfb4f71a25bc43b3c43d5a39a4fbe5e36a7a33d867b\": container with ID starting with e63a0b66f917696b8b700cfb4f71a25bc43b3c43d5a39a4fbe5e36a7a33d867b not found: ID does not exist" Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.883659 4998 scope.go:117] "RemoveContainer" containerID="251c17f3e990734bc0e154eaaf214c1f5d15b8db902b3a9dc2903061c48c64c3" Dec 03 17:00:28 crc kubenswrapper[4998]: E1203 17:00:28.883923 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"251c17f3e990734bc0e154eaaf214c1f5d15b8db902b3a9dc2903061c48c64c3\": container with ID starting with 251c17f3e990734bc0e154eaaf214c1f5d15b8db902b3a9dc2903061c48c64c3 not found: ID does not exist" containerID="251c17f3e990734bc0e154eaaf214c1f5d15b8db902b3a9dc2903061c48c64c3" Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.883939 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"251c17f3e990734bc0e154eaaf214c1f5d15b8db902b3a9dc2903061c48c64c3"} err="failed to get container status \"251c17f3e990734bc0e154eaaf214c1f5d15b8db902b3a9dc2903061c48c64c3\": rpc error: code = NotFound desc = could not find container \"251c17f3e990734bc0e154eaaf214c1f5d15b8db902b3a9dc2903061c48c64c3\": container with ID starting with 251c17f3e990734bc0e154eaaf214c1f5d15b8db902b3a9dc2903061c48c64c3 not found: ID does not exist" Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.883950 4998 scope.go:117] "RemoveContainer" containerID="adadcb4a6a35fa670d260da850a459953da744421077c24879c8a32cf37f0409" Dec 03 17:00:28 crc kubenswrapper[4998]: E1203 17:00:28.884097 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adadcb4a6a35fa670d260da850a459953da744421077c24879c8a32cf37f0409\": container with ID starting with adadcb4a6a35fa670d260da850a459953da744421077c24879c8a32cf37f0409 not found: ID does not exist" containerID="adadcb4a6a35fa670d260da850a459953da744421077c24879c8a32cf37f0409" Dec 03 17:00:28 crc kubenswrapper[4998]: I1203 17:00:28.884113 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adadcb4a6a35fa670d260da850a459953da744421077c24879c8a32cf37f0409"} err="failed to get container status \"adadcb4a6a35fa670d260da850a459953da744421077c24879c8a32cf37f0409\": rpc error: code = NotFound desc = could not find container \"adadcb4a6a35fa670d260da850a459953da744421077c24879c8a32cf37f0409\": container with ID starting with adadcb4a6a35fa670d260da850a459953da744421077c24879c8a32cf37f0409 not found: ID does not exist" Dec 03 17:00:29 crc kubenswrapper[4998]: I1203 17:00:29.690059 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="816953d5-a5ca-4822-94cb-600d1bbbdb6b" path="/var/lib/kubelet/pods/816953d5-a5ca-4822-94cb-600d1bbbdb6b/volumes" Dec 03 17:01:00 crc kubenswrapper[4998]: I1203 17:01:00.160975 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29413021-8jrpv"] Dec 03 17:01:00 crc kubenswrapper[4998]: E1203 17:01:00.162310 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="816953d5-a5ca-4822-94cb-600d1bbbdb6b" containerName="registry-server" Dec 03 17:01:00 crc kubenswrapper[4998]: I1203 17:01:00.162335 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="816953d5-a5ca-4822-94cb-600d1bbbdb6b" containerName="registry-server" Dec 03 17:01:00 crc kubenswrapper[4998]: E1203 17:01:00.162375 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="816953d5-a5ca-4822-94cb-600d1bbbdb6b" containerName="extract-content" Dec 03 17:01:00 crc kubenswrapper[4998]: I1203 17:01:00.162389 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="816953d5-a5ca-4822-94cb-600d1bbbdb6b" containerName="extract-content" Dec 03 17:01:00 crc kubenswrapper[4998]: E1203 17:01:00.162436 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="816953d5-a5ca-4822-94cb-600d1bbbdb6b" containerName="extract-utilities" Dec 03 17:01:00 crc kubenswrapper[4998]: I1203 17:01:00.162450 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="816953d5-a5ca-4822-94cb-600d1bbbdb6b" containerName="extract-utilities" Dec 03 17:01:00 crc kubenswrapper[4998]: I1203 17:01:00.162889 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="816953d5-a5ca-4822-94cb-600d1bbbdb6b" containerName="registry-server" Dec 03 17:01:00 crc kubenswrapper[4998]: I1203 17:01:00.164164 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413021-8jrpv" Dec 03 17:01:00 crc kubenswrapper[4998]: I1203 17:01:00.172043 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29413021-8jrpv"] Dec 03 17:01:00 crc kubenswrapper[4998]: I1203 17:01:00.198232 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5swbg\" (UniqueName: \"kubernetes.io/projected/526572d2-2bee-415c-a31c-def5dd4f520a-kube-api-access-5swbg\") pod \"keystone-cron-29413021-8jrpv\" (UID: \"526572d2-2bee-415c-a31c-def5dd4f520a\") " pod="openstack/keystone-cron-29413021-8jrpv" Dec 03 17:01:00 crc kubenswrapper[4998]: I1203 17:01:00.198312 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/526572d2-2bee-415c-a31c-def5dd4f520a-fernet-keys\") pod \"keystone-cron-29413021-8jrpv\" (UID: \"526572d2-2bee-415c-a31c-def5dd4f520a\") " pod="openstack/keystone-cron-29413021-8jrpv" Dec 03 17:01:00 crc kubenswrapper[4998]: I1203 17:01:00.198448 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/526572d2-2bee-415c-a31c-def5dd4f520a-combined-ca-bundle\") pod \"keystone-cron-29413021-8jrpv\" (UID: \"526572d2-2bee-415c-a31c-def5dd4f520a\") " pod="openstack/keystone-cron-29413021-8jrpv" Dec 03 17:01:00 crc kubenswrapper[4998]: I1203 17:01:00.198502 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/526572d2-2bee-415c-a31c-def5dd4f520a-config-data\") pod \"keystone-cron-29413021-8jrpv\" (UID: \"526572d2-2bee-415c-a31c-def5dd4f520a\") " pod="openstack/keystone-cron-29413021-8jrpv" Dec 03 17:01:00 crc kubenswrapper[4998]: I1203 17:01:00.300428 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/526572d2-2bee-415c-a31c-def5dd4f520a-combined-ca-bundle\") pod \"keystone-cron-29413021-8jrpv\" (UID: \"526572d2-2bee-415c-a31c-def5dd4f520a\") " pod="openstack/keystone-cron-29413021-8jrpv" Dec 03 17:01:00 crc kubenswrapper[4998]: I1203 17:01:00.300559 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/526572d2-2bee-415c-a31c-def5dd4f520a-config-data\") pod \"keystone-cron-29413021-8jrpv\" (UID: \"526572d2-2bee-415c-a31c-def5dd4f520a\") " pod="openstack/keystone-cron-29413021-8jrpv" Dec 03 17:01:00 crc kubenswrapper[4998]: I1203 17:01:00.300676 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5swbg\" (UniqueName: \"kubernetes.io/projected/526572d2-2bee-415c-a31c-def5dd4f520a-kube-api-access-5swbg\") pod \"keystone-cron-29413021-8jrpv\" (UID: \"526572d2-2bee-415c-a31c-def5dd4f520a\") " pod="openstack/keystone-cron-29413021-8jrpv" Dec 03 17:01:00 crc kubenswrapper[4998]: I1203 17:01:00.300740 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/526572d2-2bee-415c-a31c-def5dd4f520a-fernet-keys\") pod \"keystone-cron-29413021-8jrpv\" (UID: \"526572d2-2bee-415c-a31c-def5dd4f520a\") " pod="openstack/keystone-cron-29413021-8jrpv" Dec 03 17:01:00 crc kubenswrapper[4998]: I1203 17:01:00.307381 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/526572d2-2bee-415c-a31c-def5dd4f520a-combined-ca-bundle\") pod \"keystone-cron-29413021-8jrpv\" (UID: \"526572d2-2bee-415c-a31c-def5dd4f520a\") " pod="openstack/keystone-cron-29413021-8jrpv" Dec 03 17:01:00 crc kubenswrapper[4998]: I1203 17:01:00.307676 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/526572d2-2bee-415c-a31c-def5dd4f520a-config-data\") pod \"keystone-cron-29413021-8jrpv\" (UID: \"526572d2-2bee-415c-a31c-def5dd4f520a\") " pod="openstack/keystone-cron-29413021-8jrpv" Dec 03 17:01:00 crc kubenswrapper[4998]: I1203 17:01:00.311639 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/526572d2-2bee-415c-a31c-def5dd4f520a-fernet-keys\") pod \"keystone-cron-29413021-8jrpv\" (UID: \"526572d2-2bee-415c-a31c-def5dd4f520a\") " pod="openstack/keystone-cron-29413021-8jrpv" Dec 03 17:01:00 crc kubenswrapper[4998]: I1203 17:01:00.323281 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5swbg\" (UniqueName: \"kubernetes.io/projected/526572d2-2bee-415c-a31c-def5dd4f520a-kube-api-access-5swbg\") pod \"keystone-cron-29413021-8jrpv\" (UID: \"526572d2-2bee-415c-a31c-def5dd4f520a\") " pod="openstack/keystone-cron-29413021-8jrpv" Dec 03 17:01:00 crc kubenswrapper[4998]: I1203 17:01:00.517463 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413021-8jrpv" Dec 03 17:01:00 crc kubenswrapper[4998]: I1203 17:01:00.973029 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29413021-8jrpv"] Dec 03 17:01:01 crc kubenswrapper[4998]: I1203 17:01:01.131002 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413021-8jrpv" event={"ID":"526572d2-2bee-415c-a31c-def5dd4f520a","Type":"ContainerStarted","Data":"dbeb3eb08ac228cddd2ec93bf5910dfe6d1799a48a540e5f9978c130047fc238"} Dec 03 17:01:01 crc kubenswrapper[4998]: I1203 17:01:01.340674 4998 scope.go:117] "RemoveContainer" containerID="d82d3f9946783abd8003d19ede7c6ae7f58f97e12e99fa71b406eb485cb1eb8e" Dec 03 17:01:02 crc kubenswrapper[4998]: I1203 17:01:02.146552 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413021-8jrpv" event={"ID":"526572d2-2bee-415c-a31c-def5dd4f520a","Type":"ContainerStarted","Data":"4b11f1654a8c3e4fb66ee607e20bb8787aff0889eefa6222fb6d643c2214d7ea"} Dec 03 17:01:02 crc kubenswrapper[4998]: I1203 17:01:02.173649 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29413021-8jrpv" podStartSLOduration=2.17361742 podStartE2EDuration="2.17361742s" podCreationTimestamp="2025-12-03 17:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:01:02.170539152 +0000 UTC m=+3440.782239415" watchObservedRunningTime="2025-12-03 17:01:02.17361742 +0000 UTC m=+3440.785317673" Dec 03 17:01:05 crc kubenswrapper[4998]: I1203 17:01:05.186187 4998 generic.go:334] "Generic (PLEG): container finished" podID="526572d2-2bee-415c-a31c-def5dd4f520a" containerID="4b11f1654a8c3e4fb66ee607e20bb8787aff0889eefa6222fb6d643c2214d7ea" exitCode=0 Dec 03 17:01:05 crc kubenswrapper[4998]: I1203 17:01:05.186243 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413021-8jrpv" event={"ID":"526572d2-2bee-415c-a31c-def5dd4f520a","Type":"ContainerDied","Data":"4b11f1654a8c3e4fb66ee607e20bb8787aff0889eefa6222fb6d643c2214d7ea"} Dec 03 17:01:06 crc kubenswrapper[4998]: I1203 17:01:06.572183 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413021-8jrpv" Dec 03 17:01:06 crc kubenswrapper[4998]: I1203 17:01:06.761324 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5swbg\" (UniqueName: \"kubernetes.io/projected/526572d2-2bee-415c-a31c-def5dd4f520a-kube-api-access-5swbg\") pod \"526572d2-2bee-415c-a31c-def5dd4f520a\" (UID: \"526572d2-2bee-415c-a31c-def5dd4f520a\") " Dec 03 17:01:06 crc kubenswrapper[4998]: I1203 17:01:06.761397 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/526572d2-2bee-415c-a31c-def5dd4f520a-fernet-keys\") pod \"526572d2-2bee-415c-a31c-def5dd4f520a\" (UID: \"526572d2-2bee-415c-a31c-def5dd4f520a\") " Dec 03 17:01:06 crc kubenswrapper[4998]: I1203 17:01:06.761439 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/526572d2-2bee-415c-a31c-def5dd4f520a-combined-ca-bundle\") pod \"526572d2-2bee-415c-a31c-def5dd4f520a\" (UID: \"526572d2-2bee-415c-a31c-def5dd4f520a\") " Dec 03 17:01:06 crc kubenswrapper[4998]: I1203 17:01:06.761572 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/526572d2-2bee-415c-a31c-def5dd4f520a-config-data\") pod \"526572d2-2bee-415c-a31c-def5dd4f520a\" (UID: \"526572d2-2bee-415c-a31c-def5dd4f520a\") " Dec 03 17:01:06 crc kubenswrapper[4998]: I1203 17:01:06.767284 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/526572d2-2bee-415c-a31c-def5dd4f520a-kube-api-access-5swbg" (OuterVolumeSpecName: "kube-api-access-5swbg") pod "526572d2-2bee-415c-a31c-def5dd4f520a" (UID: "526572d2-2bee-415c-a31c-def5dd4f520a"). InnerVolumeSpecName "kube-api-access-5swbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:01:06 crc kubenswrapper[4998]: I1203 17:01:06.771029 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/526572d2-2bee-415c-a31c-def5dd4f520a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "526572d2-2bee-415c-a31c-def5dd4f520a" (UID: "526572d2-2bee-415c-a31c-def5dd4f520a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:01:06 crc kubenswrapper[4998]: I1203 17:01:06.792370 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/526572d2-2bee-415c-a31c-def5dd4f520a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "526572d2-2bee-415c-a31c-def5dd4f520a" (UID: "526572d2-2bee-415c-a31c-def5dd4f520a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:01:06 crc kubenswrapper[4998]: I1203 17:01:06.830700 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/526572d2-2bee-415c-a31c-def5dd4f520a-config-data" (OuterVolumeSpecName: "config-data") pod "526572d2-2bee-415c-a31c-def5dd4f520a" (UID: "526572d2-2bee-415c-a31c-def5dd4f520a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:01:06 crc kubenswrapper[4998]: I1203 17:01:06.864647 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5swbg\" (UniqueName: \"kubernetes.io/projected/526572d2-2bee-415c-a31c-def5dd4f520a-kube-api-access-5swbg\") on node \"crc\" DevicePath \"\"" Dec 03 17:01:06 crc kubenswrapper[4998]: I1203 17:01:06.864689 4998 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/526572d2-2bee-415c-a31c-def5dd4f520a-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 17:01:06 crc kubenswrapper[4998]: I1203 17:01:06.864707 4998 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/526572d2-2bee-415c-a31c-def5dd4f520a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:01:06 crc kubenswrapper[4998]: I1203 17:01:06.864726 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/526572d2-2bee-415c-a31c-def5dd4f520a-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:01:07 crc kubenswrapper[4998]: I1203 17:01:07.205376 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413021-8jrpv" event={"ID":"526572d2-2bee-415c-a31c-def5dd4f520a","Type":"ContainerDied","Data":"dbeb3eb08ac228cddd2ec93bf5910dfe6d1799a48a540e5f9978c130047fc238"} Dec 03 17:01:07 crc kubenswrapper[4998]: I1203 17:01:07.205414 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbeb3eb08ac228cddd2ec93bf5910dfe6d1799a48a540e5f9978c130047fc238" Dec 03 17:01:07 crc kubenswrapper[4998]: I1203 17:01:07.205475 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413021-8jrpv" Dec 03 17:01:57 crc kubenswrapper[4998]: I1203 17:01:57.110967 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:01:57 crc kubenswrapper[4998]: I1203 17:01:57.111427 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:02:27 crc kubenswrapper[4998]: I1203 17:02:27.111702 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:02:27 crc kubenswrapper[4998]: I1203 17:02:27.112311 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:02:57 crc kubenswrapper[4998]: I1203 17:02:57.111443 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:02:57 crc kubenswrapper[4998]: I1203 17:02:57.112299 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:02:57 crc kubenswrapper[4998]: I1203 17:02:57.112380 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 17:02:57 crc kubenswrapper[4998]: I1203 17:02:57.113578 4998 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14"} pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:02:57 crc kubenswrapper[4998]: I1203 17:02:57.113685 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" containerID="cri-o://672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" gracePeriod=600 Dec 03 17:02:58 crc kubenswrapper[4998]: E1203 17:02:58.007847 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:02:58 crc kubenswrapper[4998]: I1203 17:02:58.424104 4998 generic.go:334] "Generic (PLEG): container finished" podID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" exitCode=0 Dec 03 17:02:58 crc kubenswrapper[4998]: I1203 17:02:58.424986 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerDied","Data":"672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14"} Dec 03 17:02:58 crc kubenswrapper[4998]: I1203 17:02:58.425075 4998 scope.go:117] "RemoveContainer" containerID="84aefa566144217de53fa4d3aefaa1bc2dcd21fcd378c021e19510c75d4d205e" Dec 03 17:02:58 crc kubenswrapper[4998]: I1203 17:02:58.426010 4998 scope.go:117] "RemoveContainer" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" Dec 03 17:02:58 crc kubenswrapper[4998]: E1203 17:02:58.426530 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:03:11 crc kubenswrapper[4998]: I1203 17:03:11.686007 4998 scope.go:117] "RemoveContainer" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" Dec 03 17:03:11 crc kubenswrapper[4998]: E1203 17:03:11.687004 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:03:17 crc kubenswrapper[4998]: I1203 17:03:17.055926 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kst4v"] Dec 03 17:03:17 crc kubenswrapper[4998]: E1203 17:03:17.056947 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="526572d2-2bee-415c-a31c-def5dd4f520a" containerName="keystone-cron" Dec 03 17:03:17 crc kubenswrapper[4998]: I1203 17:03:17.056961 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="526572d2-2bee-415c-a31c-def5dd4f520a" containerName="keystone-cron" Dec 03 17:03:17 crc kubenswrapper[4998]: I1203 17:03:17.057161 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="526572d2-2bee-415c-a31c-def5dd4f520a" containerName="keystone-cron" Dec 03 17:03:17 crc kubenswrapper[4998]: I1203 17:03:17.062686 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kst4v" Dec 03 17:03:17 crc kubenswrapper[4998]: I1203 17:03:17.070985 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kst4v"] Dec 03 17:03:17 crc kubenswrapper[4998]: I1203 17:03:17.169705 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06c5576c-c545-4798-86e5-3e82c1df566e-catalog-content\") pod \"certified-operators-kst4v\" (UID: \"06c5576c-c545-4798-86e5-3e82c1df566e\") " pod="openshift-marketplace/certified-operators-kst4v" Dec 03 17:03:17 crc kubenswrapper[4998]: I1203 17:03:17.169776 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84h6q\" (UniqueName: \"kubernetes.io/projected/06c5576c-c545-4798-86e5-3e82c1df566e-kube-api-access-84h6q\") pod \"certified-operators-kst4v\" (UID: \"06c5576c-c545-4798-86e5-3e82c1df566e\") " pod="openshift-marketplace/certified-operators-kst4v" Dec 03 17:03:17 crc kubenswrapper[4998]: I1203 17:03:17.169878 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06c5576c-c545-4798-86e5-3e82c1df566e-utilities\") pod \"certified-operators-kst4v\" (UID: \"06c5576c-c545-4798-86e5-3e82c1df566e\") " pod="openshift-marketplace/certified-operators-kst4v" Dec 03 17:03:17 crc kubenswrapper[4998]: I1203 17:03:17.271906 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06c5576c-c545-4798-86e5-3e82c1df566e-utilities\") pod \"certified-operators-kst4v\" (UID: \"06c5576c-c545-4798-86e5-3e82c1df566e\") " pod="openshift-marketplace/certified-operators-kst4v" Dec 03 17:03:17 crc kubenswrapper[4998]: I1203 17:03:17.272090 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06c5576c-c545-4798-86e5-3e82c1df566e-catalog-content\") pod \"certified-operators-kst4v\" (UID: \"06c5576c-c545-4798-86e5-3e82c1df566e\") " pod="openshift-marketplace/certified-operators-kst4v" Dec 03 17:03:17 crc kubenswrapper[4998]: I1203 17:03:17.272131 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84h6q\" (UniqueName: \"kubernetes.io/projected/06c5576c-c545-4798-86e5-3e82c1df566e-kube-api-access-84h6q\") pod \"certified-operators-kst4v\" (UID: \"06c5576c-c545-4798-86e5-3e82c1df566e\") " pod="openshift-marketplace/certified-operators-kst4v" Dec 03 17:03:17 crc kubenswrapper[4998]: I1203 17:03:17.272435 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06c5576c-c545-4798-86e5-3e82c1df566e-utilities\") pod \"certified-operators-kst4v\" (UID: \"06c5576c-c545-4798-86e5-3e82c1df566e\") " pod="openshift-marketplace/certified-operators-kst4v" Dec 03 17:03:17 crc kubenswrapper[4998]: I1203 17:03:17.272720 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06c5576c-c545-4798-86e5-3e82c1df566e-catalog-content\") pod \"certified-operators-kst4v\" (UID: \"06c5576c-c545-4798-86e5-3e82c1df566e\") " pod="openshift-marketplace/certified-operators-kst4v" Dec 03 17:03:17 crc kubenswrapper[4998]: I1203 17:03:17.301880 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84h6q\" (UniqueName: \"kubernetes.io/projected/06c5576c-c545-4798-86e5-3e82c1df566e-kube-api-access-84h6q\") pod \"certified-operators-kst4v\" (UID: \"06c5576c-c545-4798-86e5-3e82c1df566e\") " pod="openshift-marketplace/certified-operators-kst4v" Dec 03 17:03:17 crc kubenswrapper[4998]: I1203 17:03:17.385361 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kst4v" Dec 03 17:03:18 crc kubenswrapper[4998]: I1203 17:03:17.995539 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kst4v"] Dec 03 17:03:18 crc kubenswrapper[4998]: I1203 17:03:18.639656 4998 generic.go:334] "Generic (PLEG): container finished" podID="06c5576c-c545-4798-86e5-3e82c1df566e" containerID="6720bb243cb14b757f31d6545c02411c4eb98ad8296307ab4443987f853ae21d" exitCode=0 Dec 03 17:03:18 crc kubenswrapper[4998]: I1203 17:03:18.639705 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kst4v" event={"ID":"06c5576c-c545-4798-86e5-3e82c1df566e","Type":"ContainerDied","Data":"6720bb243cb14b757f31d6545c02411c4eb98ad8296307ab4443987f853ae21d"} Dec 03 17:03:18 crc kubenswrapper[4998]: I1203 17:03:18.639730 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kst4v" event={"ID":"06c5576c-c545-4798-86e5-3e82c1df566e","Type":"ContainerStarted","Data":"51a05f6b2ec393aa2eadd7dd55dcf0c546fac11cf7c7e39b006380e68537cc76"} Dec 03 17:03:20 crc kubenswrapper[4998]: I1203 17:03:20.665764 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kst4v" event={"ID":"06c5576c-c545-4798-86e5-3e82c1df566e","Type":"ContainerStarted","Data":"dff1cebba7d30dc9bee965651f328758f30683402e136d6a35f529cd6b89f4fc"} Dec 03 17:03:25 crc kubenswrapper[4998]: I1203 17:03:24.678174 4998 scope.go:117] "RemoveContainer" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" Dec 03 17:03:25 crc kubenswrapper[4998]: E1203 17:03:24.678814 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:03:25 crc kubenswrapper[4998]: I1203 17:03:25.713128 4998 generic.go:334] "Generic (PLEG): container finished" podID="06c5576c-c545-4798-86e5-3e82c1df566e" containerID="dff1cebba7d30dc9bee965651f328758f30683402e136d6a35f529cd6b89f4fc" exitCode=0 Dec 03 17:03:25 crc kubenswrapper[4998]: I1203 17:03:25.713245 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kst4v" event={"ID":"06c5576c-c545-4798-86e5-3e82c1df566e","Type":"ContainerDied","Data":"dff1cebba7d30dc9bee965651f328758f30683402e136d6a35f529cd6b89f4fc"} Dec 03 17:03:27 crc kubenswrapper[4998]: I1203 17:03:27.738632 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kst4v" event={"ID":"06c5576c-c545-4798-86e5-3e82c1df566e","Type":"ContainerStarted","Data":"58e877b37ed009151732b887b7b26c577a460a0f2adc15911ab29b8c2551687d"} Dec 03 17:03:28 crc kubenswrapper[4998]: I1203 17:03:28.782076 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kst4v" podStartSLOduration=3.322809628 podStartE2EDuration="11.782058373s" podCreationTimestamp="2025-12-03 17:03:17 +0000 UTC" firstStartedPulling="2025-12-03 17:03:18.64139577 +0000 UTC m=+3577.253095993" lastFinishedPulling="2025-12-03 17:03:27.100644515 +0000 UTC m=+3585.712344738" observedRunningTime="2025-12-03 17:03:28.773459304 +0000 UTC m=+3587.385159527" watchObservedRunningTime="2025-12-03 17:03:28.782058373 +0000 UTC m=+3587.393758596" Dec 03 17:03:36 crc kubenswrapper[4998]: I1203 17:03:36.739089 4998 scope.go:117] "RemoveContainer" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" Dec 03 17:03:36 crc kubenswrapper[4998]: E1203 17:03:36.740477 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:03:37 crc kubenswrapper[4998]: I1203 17:03:37.386225 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kst4v" Dec 03 17:03:37 crc kubenswrapper[4998]: I1203 17:03:37.386284 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kst4v" Dec 03 17:03:37 crc kubenswrapper[4998]: I1203 17:03:37.440847 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kst4v" Dec 03 17:03:37 crc kubenswrapper[4998]: I1203 17:03:37.906233 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kst4v" Dec 03 17:03:37 crc kubenswrapper[4998]: I1203 17:03:37.961707 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kst4v"] Dec 03 17:03:39 crc kubenswrapper[4998]: I1203 17:03:39.868773 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kst4v" podUID="06c5576c-c545-4798-86e5-3e82c1df566e" containerName="registry-server" containerID="cri-o://58e877b37ed009151732b887b7b26c577a460a0f2adc15911ab29b8c2551687d" gracePeriod=2 Dec 03 17:03:40 crc kubenswrapper[4998]: I1203 17:03:40.441074 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kst4v" Dec 03 17:03:40 crc kubenswrapper[4998]: I1203 17:03:40.525200 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84h6q\" (UniqueName: \"kubernetes.io/projected/06c5576c-c545-4798-86e5-3e82c1df566e-kube-api-access-84h6q\") pod \"06c5576c-c545-4798-86e5-3e82c1df566e\" (UID: \"06c5576c-c545-4798-86e5-3e82c1df566e\") " Dec 03 17:03:40 crc kubenswrapper[4998]: I1203 17:03:40.525544 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06c5576c-c545-4798-86e5-3e82c1df566e-utilities\") pod \"06c5576c-c545-4798-86e5-3e82c1df566e\" (UID: \"06c5576c-c545-4798-86e5-3e82c1df566e\") " Dec 03 17:03:40 crc kubenswrapper[4998]: I1203 17:03:40.525840 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06c5576c-c545-4798-86e5-3e82c1df566e-catalog-content\") pod \"06c5576c-c545-4798-86e5-3e82c1df566e\" (UID: \"06c5576c-c545-4798-86e5-3e82c1df566e\") " Dec 03 17:03:40 crc kubenswrapper[4998]: I1203 17:03:40.526533 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06c5576c-c545-4798-86e5-3e82c1df566e-utilities" (OuterVolumeSpecName: "utilities") pod "06c5576c-c545-4798-86e5-3e82c1df566e" (UID: "06c5576c-c545-4798-86e5-3e82c1df566e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:03:40 crc kubenswrapper[4998]: I1203 17:03:40.526839 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06c5576c-c545-4798-86e5-3e82c1df566e-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:03:40 crc kubenswrapper[4998]: I1203 17:03:40.531789 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06c5576c-c545-4798-86e5-3e82c1df566e-kube-api-access-84h6q" (OuterVolumeSpecName: "kube-api-access-84h6q") pod "06c5576c-c545-4798-86e5-3e82c1df566e" (UID: "06c5576c-c545-4798-86e5-3e82c1df566e"). InnerVolumeSpecName "kube-api-access-84h6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:03:40 crc kubenswrapper[4998]: I1203 17:03:40.597936 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06c5576c-c545-4798-86e5-3e82c1df566e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "06c5576c-c545-4798-86e5-3e82c1df566e" (UID: "06c5576c-c545-4798-86e5-3e82c1df566e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:03:40 crc kubenswrapper[4998]: I1203 17:03:40.628615 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06c5576c-c545-4798-86e5-3e82c1df566e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:03:40 crc kubenswrapper[4998]: I1203 17:03:40.628657 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84h6q\" (UniqueName: \"kubernetes.io/projected/06c5576c-c545-4798-86e5-3e82c1df566e-kube-api-access-84h6q\") on node \"crc\" DevicePath \"\"" Dec 03 17:03:40 crc kubenswrapper[4998]: I1203 17:03:40.879368 4998 generic.go:334] "Generic (PLEG): container finished" podID="06c5576c-c545-4798-86e5-3e82c1df566e" containerID="58e877b37ed009151732b887b7b26c577a460a0f2adc15911ab29b8c2551687d" exitCode=0 Dec 03 17:03:40 crc kubenswrapper[4998]: I1203 17:03:40.879462 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kst4v" event={"ID":"06c5576c-c545-4798-86e5-3e82c1df566e","Type":"ContainerDied","Data":"58e877b37ed009151732b887b7b26c577a460a0f2adc15911ab29b8c2551687d"} Dec 03 17:03:40 crc kubenswrapper[4998]: I1203 17:03:40.880031 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kst4v" event={"ID":"06c5576c-c545-4798-86e5-3e82c1df566e","Type":"ContainerDied","Data":"51a05f6b2ec393aa2eadd7dd55dcf0c546fac11cf7c7e39b006380e68537cc76"} Dec 03 17:03:40 crc kubenswrapper[4998]: I1203 17:03:40.880055 4998 scope.go:117] "RemoveContainer" containerID="58e877b37ed009151732b887b7b26c577a460a0f2adc15911ab29b8c2551687d" Dec 03 17:03:40 crc kubenswrapper[4998]: I1203 17:03:40.879468 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kst4v" Dec 03 17:03:40 crc kubenswrapper[4998]: I1203 17:03:40.900338 4998 scope.go:117] "RemoveContainer" containerID="dff1cebba7d30dc9bee965651f328758f30683402e136d6a35f529cd6b89f4fc" Dec 03 17:03:40 crc kubenswrapper[4998]: I1203 17:03:40.915190 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kst4v"] Dec 03 17:03:40 crc kubenswrapper[4998]: I1203 17:03:40.925249 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kst4v"] Dec 03 17:03:40 crc kubenswrapper[4998]: I1203 17:03:40.945389 4998 scope.go:117] "RemoveContainer" containerID="6720bb243cb14b757f31d6545c02411c4eb98ad8296307ab4443987f853ae21d" Dec 03 17:03:40 crc kubenswrapper[4998]: I1203 17:03:40.972514 4998 scope.go:117] "RemoveContainer" containerID="58e877b37ed009151732b887b7b26c577a460a0f2adc15911ab29b8c2551687d" Dec 03 17:03:40 crc kubenswrapper[4998]: E1203 17:03:40.973781 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58e877b37ed009151732b887b7b26c577a460a0f2adc15911ab29b8c2551687d\": container with ID starting with 58e877b37ed009151732b887b7b26c577a460a0f2adc15911ab29b8c2551687d not found: ID does not exist" containerID="58e877b37ed009151732b887b7b26c577a460a0f2adc15911ab29b8c2551687d" Dec 03 17:03:40 crc kubenswrapper[4998]: I1203 17:03:40.973827 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58e877b37ed009151732b887b7b26c577a460a0f2adc15911ab29b8c2551687d"} err="failed to get container status \"58e877b37ed009151732b887b7b26c577a460a0f2adc15911ab29b8c2551687d\": rpc error: code = NotFound desc = could not find container \"58e877b37ed009151732b887b7b26c577a460a0f2adc15911ab29b8c2551687d\": container with ID starting with 58e877b37ed009151732b887b7b26c577a460a0f2adc15911ab29b8c2551687d not found: ID does not exist" Dec 03 17:03:40 crc kubenswrapper[4998]: I1203 17:03:40.973871 4998 scope.go:117] "RemoveContainer" containerID="dff1cebba7d30dc9bee965651f328758f30683402e136d6a35f529cd6b89f4fc" Dec 03 17:03:40 crc kubenswrapper[4998]: E1203 17:03:40.974520 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dff1cebba7d30dc9bee965651f328758f30683402e136d6a35f529cd6b89f4fc\": container with ID starting with dff1cebba7d30dc9bee965651f328758f30683402e136d6a35f529cd6b89f4fc not found: ID does not exist" containerID="dff1cebba7d30dc9bee965651f328758f30683402e136d6a35f529cd6b89f4fc" Dec 03 17:03:40 crc kubenswrapper[4998]: I1203 17:03:40.974577 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dff1cebba7d30dc9bee965651f328758f30683402e136d6a35f529cd6b89f4fc"} err="failed to get container status \"dff1cebba7d30dc9bee965651f328758f30683402e136d6a35f529cd6b89f4fc\": rpc error: code = NotFound desc = could not find container \"dff1cebba7d30dc9bee965651f328758f30683402e136d6a35f529cd6b89f4fc\": container with ID starting with dff1cebba7d30dc9bee965651f328758f30683402e136d6a35f529cd6b89f4fc not found: ID does not exist" Dec 03 17:03:40 crc kubenswrapper[4998]: I1203 17:03:40.974608 4998 scope.go:117] "RemoveContainer" containerID="6720bb243cb14b757f31d6545c02411c4eb98ad8296307ab4443987f853ae21d" Dec 03 17:03:40 crc kubenswrapper[4998]: E1203 17:03:40.975077 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6720bb243cb14b757f31d6545c02411c4eb98ad8296307ab4443987f853ae21d\": container with ID starting with 6720bb243cb14b757f31d6545c02411c4eb98ad8296307ab4443987f853ae21d not found: ID does not exist" containerID="6720bb243cb14b757f31d6545c02411c4eb98ad8296307ab4443987f853ae21d" Dec 03 17:03:40 crc kubenswrapper[4998]: I1203 17:03:40.975124 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6720bb243cb14b757f31d6545c02411c4eb98ad8296307ab4443987f853ae21d"} err="failed to get container status \"6720bb243cb14b757f31d6545c02411c4eb98ad8296307ab4443987f853ae21d\": rpc error: code = NotFound desc = could not find container \"6720bb243cb14b757f31d6545c02411c4eb98ad8296307ab4443987f853ae21d\": container with ID starting with 6720bb243cb14b757f31d6545c02411c4eb98ad8296307ab4443987f853ae21d not found: ID does not exist" Dec 03 17:03:41 crc kubenswrapper[4998]: I1203 17:03:41.690935 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06c5576c-c545-4798-86e5-3e82c1df566e" path="/var/lib/kubelet/pods/06c5576c-c545-4798-86e5-3e82c1df566e/volumes" Dec 03 17:03:47 crc kubenswrapper[4998]: I1203 17:03:47.678040 4998 scope.go:117] "RemoveContainer" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" Dec 03 17:03:47 crc kubenswrapper[4998]: E1203 17:03:47.679065 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:04:00 crc kubenswrapper[4998]: I1203 17:04:00.678191 4998 scope.go:117] "RemoveContainer" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" Dec 03 17:04:00 crc kubenswrapper[4998]: E1203 17:04:00.679046 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:04:12 crc kubenswrapper[4998]: I1203 17:04:12.679038 4998 scope.go:117] "RemoveContainer" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" Dec 03 17:04:12 crc kubenswrapper[4998]: E1203 17:04:12.680137 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:04:23 crc kubenswrapper[4998]: I1203 17:04:23.678555 4998 scope.go:117] "RemoveContainer" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" Dec 03 17:04:23 crc kubenswrapper[4998]: E1203 17:04:23.679539 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:04:37 crc kubenswrapper[4998]: I1203 17:04:37.678982 4998 scope.go:117] "RemoveContainer" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" Dec 03 17:04:37 crc kubenswrapper[4998]: E1203 17:04:37.680206 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:04:52 crc kubenswrapper[4998]: I1203 17:04:52.678304 4998 scope.go:117] "RemoveContainer" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" Dec 03 17:04:52 crc kubenswrapper[4998]: E1203 17:04:52.679096 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:05:06 crc kubenswrapper[4998]: I1203 17:05:06.677693 4998 scope.go:117] "RemoveContainer" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" Dec 03 17:05:06 crc kubenswrapper[4998]: E1203 17:05:06.678427 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:05:17 crc kubenswrapper[4998]: I1203 17:05:17.678949 4998 scope.go:117] "RemoveContainer" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" Dec 03 17:05:17 crc kubenswrapper[4998]: E1203 17:05:17.681445 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:05:29 crc kubenswrapper[4998]: I1203 17:05:29.678533 4998 scope.go:117] "RemoveContainer" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" Dec 03 17:05:29 crc kubenswrapper[4998]: E1203 17:05:29.680594 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:05:44 crc kubenswrapper[4998]: I1203 17:05:44.677338 4998 scope.go:117] "RemoveContainer" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" Dec 03 17:05:44 crc kubenswrapper[4998]: E1203 17:05:44.678080 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:05:57 crc kubenswrapper[4998]: I1203 17:05:57.679903 4998 scope.go:117] "RemoveContainer" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" Dec 03 17:05:57 crc kubenswrapper[4998]: E1203 17:05:57.680606 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:06:09 crc kubenswrapper[4998]: I1203 17:06:09.678632 4998 scope.go:117] "RemoveContainer" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" Dec 03 17:06:09 crc kubenswrapper[4998]: E1203 17:06:09.679598 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:06:20 crc kubenswrapper[4998]: I1203 17:06:20.678573 4998 scope.go:117] "RemoveContainer" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" Dec 03 17:06:20 crc kubenswrapper[4998]: E1203 17:06:20.679568 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:06:33 crc kubenswrapper[4998]: I1203 17:06:33.678431 4998 scope.go:117] "RemoveContainer" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" Dec 03 17:06:33 crc kubenswrapper[4998]: E1203 17:06:33.679259 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:06:47 crc kubenswrapper[4998]: I1203 17:06:47.679278 4998 scope.go:117] "RemoveContainer" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" Dec 03 17:06:47 crc kubenswrapper[4998]: E1203 17:06:47.680748 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:07:02 crc kubenswrapper[4998]: I1203 17:07:02.678269 4998 scope.go:117] "RemoveContainer" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" Dec 03 17:07:02 crc kubenswrapper[4998]: E1203 17:07:02.680141 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:07:16 crc kubenswrapper[4998]: I1203 17:07:16.678719 4998 scope.go:117] "RemoveContainer" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" Dec 03 17:07:16 crc kubenswrapper[4998]: E1203 17:07:16.679630 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:07:31 crc kubenswrapper[4998]: I1203 17:07:31.678903 4998 scope.go:117] "RemoveContainer" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" Dec 03 17:07:31 crc kubenswrapper[4998]: E1203 17:07:31.679869 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:07:33 crc kubenswrapper[4998]: I1203 17:07:33.992640 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cdbps"] Dec 03 17:07:33 crc kubenswrapper[4998]: E1203 17:07:33.993699 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06c5576c-c545-4798-86e5-3e82c1df566e" containerName="registry-server" Dec 03 17:07:33 crc kubenswrapper[4998]: I1203 17:07:33.993721 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="06c5576c-c545-4798-86e5-3e82c1df566e" containerName="registry-server" Dec 03 17:07:33 crc kubenswrapper[4998]: E1203 17:07:33.993741 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06c5576c-c545-4798-86e5-3e82c1df566e" containerName="extract-content" Dec 03 17:07:33 crc kubenswrapper[4998]: I1203 17:07:33.993771 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="06c5576c-c545-4798-86e5-3e82c1df566e" containerName="extract-content" Dec 03 17:07:33 crc kubenswrapper[4998]: E1203 17:07:33.993833 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06c5576c-c545-4798-86e5-3e82c1df566e" containerName="extract-utilities" Dec 03 17:07:33 crc kubenswrapper[4998]: I1203 17:07:33.993847 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="06c5576c-c545-4798-86e5-3e82c1df566e" containerName="extract-utilities" Dec 03 17:07:33 crc kubenswrapper[4998]: I1203 17:07:33.994119 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="06c5576c-c545-4798-86e5-3e82c1df566e" containerName="registry-server" Dec 03 17:07:34 crc kubenswrapper[4998]: I1203 17:07:33.997005 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cdbps" Dec 03 17:07:34 crc kubenswrapper[4998]: I1203 17:07:34.013950 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cdbps"] Dec 03 17:07:34 crc kubenswrapper[4998]: I1203 17:07:34.025090 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05d49678-1871-4e47-a640-118177f36a29-utilities\") pod \"redhat-marketplace-cdbps\" (UID: \"05d49678-1871-4e47-a640-118177f36a29\") " pod="openshift-marketplace/redhat-marketplace-cdbps" Dec 03 17:07:34 crc kubenswrapper[4998]: I1203 17:07:34.025218 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jz5f\" (UniqueName: \"kubernetes.io/projected/05d49678-1871-4e47-a640-118177f36a29-kube-api-access-6jz5f\") pod \"redhat-marketplace-cdbps\" (UID: \"05d49678-1871-4e47-a640-118177f36a29\") " pod="openshift-marketplace/redhat-marketplace-cdbps" Dec 03 17:07:34 crc kubenswrapper[4998]: I1203 17:07:34.025294 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05d49678-1871-4e47-a640-118177f36a29-catalog-content\") pod \"redhat-marketplace-cdbps\" (UID: \"05d49678-1871-4e47-a640-118177f36a29\") " pod="openshift-marketplace/redhat-marketplace-cdbps" Dec 03 17:07:34 crc kubenswrapper[4998]: I1203 17:07:34.127566 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05d49678-1871-4e47-a640-118177f36a29-utilities\") pod \"redhat-marketplace-cdbps\" (UID: \"05d49678-1871-4e47-a640-118177f36a29\") " pod="openshift-marketplace/redhat-marketplace-cdbps" Dec 03 17:07:34 crc kubenswrapper[4998]: I1203 17:07:34.127661 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jz5f\" (UniqueName: \"kubernetes.io/projected/05d49678-1871-4e47-a640-118177f36a29-kube-api-access-6jz5f\") pod \"redhat-marketplace-cdbps\" (UID: \"05d49678-1871-4e47-a640-118177f36a29\") " pod="openshift-marketplace/redhat-marketplace-cdbps" Dec 03 17:07:34 crc kubenswrapper[4998]: I1203 17:07:34.127722 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05d49678-1871-4e47-a640-118177f36a29-catalog-content\") pod \"redhat-marketplace-cdbps\" (UID: \"05d49678-1871-4e47-a640-118177f36a29\") " pod="openshift-marketplace/redhat-marketplace-cdbps" Dec 03 17:07:34 crc kubenswrapper[4998]: I1203 17:07:34.128206 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05d49678-1871-4e47-a640-118177f36a29-utilities\") pod \"redhat-marketplace-cdbps\" (UID: \"05d49678-1871-4e47-a640-118177f36a29\") " pod="openshift-marketplace/redhat-marketplace-cdbps" Dec 03 17:07:34 crc kubenswrapper[4998]: I1203 17:07:34.128231 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05d49678-1871-4e47-a640-118177f36a29-catalog-content\") pod \"redhat-marketplace-cdbps\" (UID: \"05d49678-1871-4e47-a640-118177f36a29\") " pod="openshift-marketplace/redhat-marketplace-cdbps" Dec 03 17:07:34 crc kubenswrapper[4998]: I1203 17:07:34.590288 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jz5f\" (UniqueName: \"kubernetes.io/projected/05d49678-1871-4e47-a640-118177f36a29-kube-api-access-6jz5f\") pod \"redhat-marketplace-cdbps\" (UID: \"05d49678-1871-4e47-a640-118177f36a29\") " pod="openshift-marketplace/redhat-marketplace-cdbps" Dec 03 17:07:34 crc kubenswrapper[4998]: I1203 17:07:34.631166 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cdbps" Dec 03 17:07:35 crc kubenswrapper[4998]: I1203 17:07:35.156581 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cdbps"] Dec 03 17:07:35 crc kubenswrapper[4998]: I1203 17:07:35.523132 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cdbps" event={"ID":"05d49678-1871-4e47-a640-118177f36a29","Type":"ContainerStarted","Data":"1914d0c651ce7444ae24a88844d57f50e1430ec65601f5e0e4ea4cfe117c979d"} Dec 03 17:07:35 crc kubenswrapper[4998]: I1203 17:07:35.524031 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cdbps" event={"ID":"05d49678-1871-4e47-a640-118177f36a29","Type":"ContainerStarted","Data":"f2f628f32bed810ee56a25b83259962542188ba506bd191e72df71017e16d755"} Dec 03 17:07:36 crc kubenswrapper[4998]: I1203 17:07:36.546449 4998 generic.go:334] "Generic (PLEG): container finished" podID="05d49678-1871-4e47-a640-118177f36a29" containerID="1914d0c651ce7444ae24a88844d57f50e1430ec65601f5e0e4ea4cfe117c979d" exitCode=0 Dec 03 17:07:36 crc kubenswrapper[4998]: I1203 17:07:36.546589 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cdbps" event={"ID":"05d49678-1871-4e47-a640-118177f36a29","Type":"ContainerDied","Data":"1914d0c651ce7444ae24a88844d57f50e1430ec65601f5e0e4ea4cfe117c979d"} Dec 03 17:07:36 crc kubenswrapper[4998]: I1203 17:07:36.554168 4998 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 17:07:38 crc kubenswrapper[4998]: I1203 17:07:38.583215 4998 generic.go:334] "Generic (PLEG): container finished" podID="05d49678-1871-4e47-a640-118177f36a29" containerID="2b28aa24b3c70760b27c631aaa555597ab0c8015a082daf58d91b29f532d9c01" exitCode=0 Dec 03 17:07:38 crc kubenswrapper[4998]: I1203 17:07:38.583298 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cdbps" event={"ID":"05d49678-1871-4e47-a640-118177f36a29","Type":"ContainerDied","Data":"2b28aa24b3c70760b27c631aaa555597ab0c8015a082daf58d91b29f532d9c01"} Dec 03 17:07:40 crc kubenswrapper[4998]: I1203 17:07:40.607233 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cdbps" event={"ID":"05d49678-1871-4e47-a640-118177f36a29","Type":"ContainerStarted","Data":"d5bdb660261e036d2cdb30802bff83e28ac203a229abcbdb5bd756caa5456311"} Dec 03 17:07:40 crc kubenswrapper[4998]: I1203 17:07:40.631884 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cdbps" podStartSLOduration=4.069261184 podStartE2EDuration="7.631863826s" podCreationTimestamp="2025-12-03 17:07:33 +0000 UTC" firstStartedPulling="2025-12-03 17:07:36.552356401 +0000 UTC m=+3835.164056664" lastFinishedPulling="2025-12-03 17:07:40.114959083 +0000 UTC m=+3838.726659306" observedRunningTime="2025-12-03 17:07:40.622807451 +0000 UTC m=+3839.234507714" watchObservedRunningTime="2025-12-03 17:07:40.631863826 +0000 UTC m=+3839.243564049" Dec 03 17:07:44 crc kubenswrapper[4998]: I1203 17:07:44.632292 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cdbps" Dec 03 17:07:44 crc kubenswrapper[4998]: I1203 17:07:44.632930 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cdbps" Dec 03 17:07:44 crc kubenswrapper[4998]: I1203 17:07:44.693135 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cdbps" Dec 03 17:07:46 crc kubenswrapper[4998]: I1203 17:07:46.678158 4998 scope.go:117] "RemoveContainer" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" Dec 03 17:07:46 crc kubenswrapper[4998]: E1203 17:07:46.678730 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:07:54 crc kubenswrapper[4998]: I1203 17:07:54.688317 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cdbps" Dec 03 17:07:54 crc kubenswrapper[4998]: I1203 17:07:54.752911 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cdbps"] Dec 03 17:07:54 crc kubenswrapper[4998]: I1203 17:07:54.756484 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cdbps" podUID="05d49678-1871-4e47-a640-118177f36a29" containerName="registry-server" containerID="cri-o://d5bdb660261e036d2cdb30802bff83e28ac203a229abcbdb5bd756caa5456311" gracePeriod=2 Dec 03 17:07:55 crc kubenswrapper[4998]: I1203 17:07:55.539135 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cdbps" Dec 03 17:07:55 crc kubenswrapper[4998]: I1203 17:07:55.696615 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jz5f\" (UniqueName: \"kubernetes.io/projected/05d49678-1871-4e47-a640-118177f36a29-kube-api-access-6jz5f\") pod \"05d49678-1871-4e47-a640-118177f36a29\" (UID: \"05d49678-1871-4e47-a640-118177f36a29\") " Dec 03 17:07:55 crc kubenswrapper[4998]: I1203 17:07:55.696686 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05d49678-1871-4e47-a640-118177f36a29-utilities\") pod \"05d49678-1871-4e47-a640-118177f36a29\" (UID: \"05d49678-1871-4e47-a640-118177f36a29\") " Dec 03 17:07:55 crc kubenswrapper[4998]: I1203 17:07:55.696733 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05d49678-1871-4e47-a640-118177f36a29-catalog-content\") pod \"05d49678-1871-4e47-a640-118177f36a29\" (UID: \"05d49678-1871-4e47-a640-118177f36a29\") " Dec 03 17:07:55 crc kubenswrapper[4998]: I1203 17:07:55.700519 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05d49678-1871-4e47-a640-118177f36a29-utilities" (OuterVolumeSpecName: "utilities") pod "05d49678-1871-4e47-a640-118177f36a29" (UID: "05d49678-1871-4e47-a640-118177f36a29"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:07:55 crc kubenswrapper[4998]: I1203 17:07:55.715983 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05d49678-1871-4e47-a640-118177f36a29-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "05d49678-1871-4e47-a640-118177f36a29" (UID: "05d49678-1871-4e47-a640-118177f36a29"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:07:55 crc kubenswrapper[4998]: I1203 17:07:55.925564 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05d49678-1871-4e47-a640-118177f36a29-kube-api-access-6jz5f" (OuterVolumeSpecName: "kube-api-access-6jz5f") pod "05d49678-1871-4e47-a640-118177f36a29" (UID: "05d49678-1871-4e47-a640-118177f36a29"). InnerVolumeSpecName "kube-api-access-6jz5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:07:55 crc kubenswrapper[4998]: I1203 17:07:55.926263 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jz5f\" (UniqueName: \"kubernetes.io/projected/05d49678-1871-4e47-a640-118177f36a29-kube-api-access-6jz5f\") pod \"05d49678-1871-4e47-a640-118177f36a29\" (UID: \"05d49678-1871-4e47-a640-118177f36a29\") " Dec 03 17:07:55 crc kubenswrapper[4998]: W1203 17:07:55.927336 4998 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/05d49678-1871-4e47-a640-118177f36a29/volumes/kubernetes.io~projected/kube-api-access-6jz5f Dec 03 17:07:55 crc kubenswrapper[4998]: I1203 17:07:55.927415 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05d49678-1871-4e47-a640-118177f36a29-kube-api-access-6jz5f" (OuterVolumeSpecName: "kube-api-access-6jz5f") pod "05d49678-1871-4e47-a640-118177f36a29" (UID: "05d49678-1871-4e47-a640-118177f36a29"). InnerVolumeSpecName "kube-api-access-6jz5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:07:55 crc kubenswrapper[4998]: I1203 17:07:55.928414 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05d49678-1871-4e47-a640-118177f36a29-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:07:55 crc kubenswrapper[4998]: I1203 17:07:55.928475 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05d49678-1871-4e47-a640-118177f36a29-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:07:55 crc kubenswrapper[4998]: I1203 17:07:55.965194 4998 generic.go:334] "Generic (PLEG): container finished" podID="05d49678-1871-4e47-a640-118177f36a29" containerID="d5bdb660261e036d2cdb30802bff83e28ac203a229abcbdb5bd756caa5456311" exitCode=0 Dec 03 17:07:55 crc kubenswrapper[4998]: I1203 17:07:55.965238 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cdbps" event={"ID":"05d49678-1871-4e47-a640-118177f36a29","Type":"ContainerDied","Data":"d5bdb660261e036d2cdb30802bff83e28ac203a229abcbdb5bd756caa5456311"} Dec 03 17:07:55 crc kubenswrapper[4998]: I1203 17:07:55.965262 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cdbps" event={"ID":"05d49678-1871-4e47-a640-118177f36a29","Type":"ContainerDied","Data":"f2f628f32bed810ee56a25b83259962542188ba506bd191e72df71017e16d755"} Dec 03 17:07:55 crc kubenswrapper[4998]: I1203 17:07:55.965283 4998 scope.go:117] "RemoveContainer" containerID="d5bdb660261e036d2cdb30802bff83e28ac203a229abcbdb5bd756caa5456311" Dec 03 17:07:55 crc kubenswrapper[4998]: I1203 17:07:55.965408 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cdbps" Dec 03 17:07:56 crc kubenswrapper[4998]: I1203 17:07:56.026732 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cdbps"] Dec 03 17:07:56 crc kubenswrapper[4998]: I1203 17:07:56.030912 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jz5f\" (UniqueName: \"kubernetes.io/projected/05d49678-1871-4e47-a640-118177f36a29-kube-api-access-6jz5f\") on node \"crc\" DevicePath \"\"" Dec 03 17:07:56 crc kubenswrapper[4998]: I1203 17:07:56.033129 4998 scope.go:117] "RemoveContainer" containerID="2b28aa24b3c70760b27c631aaa555597ab0c8015a082daf58d91b29f532d9c01" Dec 03 17:07:56 crc kubenswrapper[4998]: I1203 17:07:56.035726 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cdbps"] Dec 03 17:07:56 crc kubenswrapper[4998]: I1203 17:07:56.060296 4998 scope.go:117] "RemoveContainer" containerID="1914d0c651ce7444ae24a88844d57f50e1430ec65601f5e0e4ea4cfe117c979d" Dec 03 17:07:56 crc kubenswrapper[4998]: I1203 17:07:56.111965 4998 scope.go:117] "RemoveContainer" containerID="d5bdb660261e036d2cdb30802bff83e28ac203a229abcbdb5bd756caa5456311" Dec 03 17:07:56 crc kubenswrapper[4998]: E1203 17:07:56.112457 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5bdb660261e036d2cdb30802bff83e28ac203a229abcbdb5bd756caa5456311\": container with ID starting with d5bdb660261e036d2cdb30802bff83e28ac203a229abcbdb5bd756caa5456311 not found: ID does not exist" containerID="d5bdb660261e036d2cdb30802bff83e28ac203a229abcbdb5bd756caa5456311" Dec 03 17:07:56 crc kubenswrapper[4998]: I1203 17:07:56.112493 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5bdb660261e036d2cdb30802bff83e28ac203a229abcbdb5bd756caa5456311"} err="failed to get container status \"d5bdb660261e036d2cdb30802bff83e28ac203a229abcbdb5bd756caa5456311\": rpc error: code = NotFound desc = could not find container \"d5bdb660261e036d2cdb30802bff83e28ac203a229abcbdb5bd756caa5456311\": container with ID starting with d5bdb660261e036d2cdb30802bff83e28ac203a229abcbdb5bd756caa5456311 not found: ID does not exist" Dec 03 17:07:56 crc kubenswrapper[4998]: I1203 17:07:56.112513 4998 scope.go:117] "RemoveContainer" containerID="2b28aa24b3c70760b27c631aaa555597ab0c8015a082daf58d91b29f532d9c01" Dec 03 17:07:56 crc kubenswrapper[4998]: E1203 17:07:56.115254 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b28aa24b3c70760b27c631aaa555597ab0c8015a082daf58d91b29f532d9c01\": container with ID starting with 2b28aa24b3c70760b27c631aaa555597ab0c8015a082daf58d91b29f532d9c01 not found: ID does not exist" containerID="2b28aa24b3c70760b27c631aaa555597ab0c8015a082daf58d91b29f532d9c01" Dec 03 17:07:56 crc kubenswrapper[4998]: I1203 17:07:56.115285 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b28aa24b3c70760b27c631aaa555597ab0c8015a082daf58d91b29f532d9c01"} err="failed to get container status \"2b28aa24b3c70760b27c631aaa555597ab0c8015a082daf58d91b29f532d9c01\": rpc error: code = NotFound desc = could not find container \"2b28aa24b3c70760b27c631aaa555597ab0c8015a082daf58d91b29f532d9c01\": container with ID starting with 2b28aa24b3c70760b27c631aaa555597ab0c8015a082daf58d91b29f532d9c01 not found: ID does not exist" Dec 03 17:07:56 crc kubenswrapper[4998]: I1203 17:07:56.115307 4998 scope.go:117] "RemoveContainer" containerID="1914d0c651ce7444ae24a88844d57f50e1430ec65601f5e0e4ea4cfe117c979d" Dec 03 17:07:56 crc kubenswrapper[4998]: E1203 17:07:56.115722 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1914d0c651ce7444ae24a88844d57f50e1430ec65601f5e0e4ea4cfe117c979d\": container with ID starting with 1914d0c651ce7444ae24a88844d57f50e1430ec65601f5e0e4ea4cfe117c979d not found: ID does not exist" containerID="1914d0c651ce7444ae24a88844d57f50e1430ec65601f5e0e4ea4cfe117c979d" Dec 03 17:07:56 crc kubenswrapper[4998]: I1203 17:07:56.115744 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1914d0c651ce7444ae24a88844d57f50e1430ec65601f5e0e4ea4cfe117c979d"} err="failed to get container status \"1914d0c651ce7444ae24a88844d57f50e1430ec65601f5e0e4ea4cfe117c979d\": rpc error: code = NotFound desc = could not find container \"1914d0c651ce7444ae24a88844d57f50e1430ec65601f5e0e4ea4cfe117c979d\": container with ID starting with 1914d0c651ce7444ae24a88844d57f50e1430ec65601f5e0e4ea4cfe117c979d not found: ID does not exist" Dec 03 17:07:57 crc kubenswrapper[4998]: I1203 17:07:57.693341 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05d49678-1871-4e47-a640-118177f36a29" path="/var/lib/kubelet/pods/05d49678-1871-4e47-a640-118177f36a29/volumes" Dec 03 17:07:59 crc kubenswrapper[4998]: I1203 17:07:59.678865 4998 scope.go:117] "RemoveContainer" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" Dec 03 17:08:00 crc kubenswrapper[4998]: I1203 17:08:00.005941 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"9ac7668007b277a203e4765518b470f255fddd0cb75042129bf6f6ab95cc4cd4"} Dec 03 17:10:27 crc kubenswrapper[4998]: I1203 17:10:27.110736 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:10:27 crc kubenswrapper[4998]: I1203 17:10:27.111486 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:10:57 crc kubenswrapper[4998]: I1203 17:10:57.111226 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:10:57 crc kubenswrapper[4998]: I1203 17:10:57.111877 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:11:27 crc kubenswrapper[4998]: I1203 17:11:27.111202 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:11:27 crc kubenswrapper[4998]: I1203 17:11:27.111671 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:11:27 crc kubenswrapper[4998]: I1203 17:11:27.111719 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 17:11:27 crc kubenswrapper[4998]: I1203 17:11:27.112303 4998 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9ac7668007b277a203e4765518b470f255fddd0cb75042129bf6f6ab95cc4cd4"} pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:11:27 crc kubenswrapper[4998]: I1203 17:11:27.112353 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" containerID="cri-o://9ac7668007b277a203e4765518b470f255fddd0cb75042129bf6f6ab95cc4cd4" gracePeriod=600 Dec 03 17:11:27 crc kubenswrapper[4998]: I1203 17:11:27.504426 4998 generic.go:334] "Generic (PLEG): container finished" podID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerID="9ac7668007b277a203e4765518b470f255fddd0cb75042129bf6f6ab95cc4cd4" exitCode=0 Dec 03 17:11:27 crc kubenswrapper[4998]: I1203 17:11:27.504510 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerDied","Data":"9ac7668007b277a203e4765518b470f255fddd0cb75042129bf6f6ab95cc4cd4"} Dec 03 17:11:27 crc kubenswrapper[4998]: I1203 17:11:27.504661 4998 scope.go:117] "RemoveContainer" containerID="672d631c5362e37631df3c5af3544322e54e6bc2a2a896a36f96d50b3421af14" Dec 03 17:11:28 crc kubenswrapper[4998]: I1203 17:11:28.522439 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85"} Dec 03 17:13:08 crc kubenswrapper[4998]: I1203 17:13:08.646431 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-276n4"] Dec 03 17:13:08 crc kubenswrapper[4998]: E1203 17:13:08.647627 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05d49678-1871-4e47-a640-118177f36a29" containerName="extract-utilities" Dec 03 17:13:08 crc kubenswrapper[4998]: I1203 17:13:08.647647 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="05d49678-1871-4e47-a640-118177f36a29" containerName="extract-utilities" Dec 03 17:13:08 crc kubenswrapper[4998]: E1203 17:13:08.647680 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05d49678-1871-4e47-a640-118177f36a29" containerName="extract-content" Dec 03 17:13:08 crc kubenswrapper[4998]: I1203 17:13:08.647693 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="05d49678-1871-4e47-a640-118177f36a29" containerName="extract-content" Dec 03 17:13:08 crc kubenswrapper[4998]: E1203 17:13:08.647717 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05d49678-1871-4e47-a640-118177f36a29" containerName="registry-server" Dec 03 17:13:08 crc kubenswrapper[4998]: I1203 17:13:08.647731 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="05d49678-1871-4e47-a640-118177f36a29" containerName="registry-server" Dec 03 17:13:08 crc kubenswrapper[4998]: I1203 17:13:08.648086 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="05d49678-1871-4e47-a640-118177f36a29" containerName="registry-server" Dec 03 17:13:08 crc kubenswrapper[4998]: I1203 17:13:08.651904 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-276n4" Dec 03 17:13:08 crc kubenswrapper[4998]: I1203 17:13:08.657320 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-276n4"] Dec 03 17:13:08 crc kubenswrapper[4998]: I1203 17:13:08.697743 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26568f59-e843-42cd-b87d-3da76a9d622e-utilities\") pod \"redhat-operators-276n4\" (UID: \"26568f59-e843-42cd-b87d-3da76a9d622e\") " pod="openshift-marketplace/redhat-operators-276n4" Dec 03 17:13:08 crc kubenswrapper[4998]: I1203 17:13:08.698047 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfm5s\" (UniqueName: \"kubernetes.io/projected/26568f59-e843-42cd-b87d-3da76a9d622e-kube-api-access-wfm5s\") pod \"redhat-operators-276n4\" (UID: \"26568f59-e843-42cd-b87d-3da76a9d622e\") " pod="openshift-marketplace/redhat-operators-276n4" Dec 03 17:13:08 crc kubenswrapper[4998]: I1203 17:13:08.698201 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26568f59-e843-42cd-b87d-3da76a9d622e-catalog-content\") pod \"redhat-operators-276n4\" (UID: \"26568f59-e843-42cd-b87d-3da76a9d622e\") " pod="openshift-marketplace/redhat-operators-276n4" Dec 03 17:13:08 crc kubenswrapper[4998]: I1203 17:13:08.800649 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26568f59-e843-42cd-b87d-3da76a9d622e-utilities\") pod \"redhat-operators-276n4\" (UID: \"26568f59-e843-42cd-b87d-3da76a9d622e\") " pod="openshift-marketplace/redhat-operators-276n4" Dec 03 17:13:08 crc kubenswrapper[4998]: I1203 17:13:08.800840 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfm5s\" (UniqueName: \"kubernetes.io/projected/26568f59-e843-42cd-b87d-3da76a9d622e-kube-api-access-wfm5s\") pod \"redhat-operators-276n4\" (UID: \"26568f59-e843-42cd-b87d-3da76a9d622e\") " pod="openshift-marketplace/redhat-operators-276n4" Dec 03 17:13:08 crc kubenswrapper[4998]: I1203 17:13:08.800887 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26568f59-e843-42cd-b87d-3da76a9d622e-catalog-content\") pod \"redhat-operators-276n4\" (UID: \"26568f59-e843-42cd-b87d-3da76a9d622e\") " pod="openshift-marketplace/redhat-operators-276n4" Dec 03 17:13:08 crc kubenswrapper[4998]: I1203 17:13:08.801233 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26568f59-e843-42cd-b87d-3da76a9d622e-utilities\") pod \"redhat-operators-276n4\" (UID: \"26568f59-e843-42cd-b87d-3da76a9d622e\") " pod="openshift-marketplace/redhat-operators-276n4" Dec 03 17:13:08 crc kubenswrapper[4998]: I1203 17:13:08.801434 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26568f59-e843-42cd-b87d-3da76a9d622e-catalog-content\") pod \"redhat-operators-276n4\" (UID: \"26568f59-e843-42cd-b87d-3da76a9d622e\") " pod="openshift-marketplace/redhat-operators-276n4" Dec 03 17:13:09 crc kubenswrapper[4998]: I1203 17:13:09.088943 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfm5s\" (UniqueName: \"kubernetes.io/projected/26568f59-e843-42cd-b87d-3da76a9d622e-kube-api-access-wfm5s\") pod \"redhat-operators-276n4\" (UID: \"26568f59-e843-42cd-b87d-3da76a9d622e\") " pod="openshift-marketplace/redhat-operators-276n4" Dec 03 17:13:09 crc kubenswrapper[4998]: I1203 17:13:09.296505 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-276n4" Dec 03 17:13:09 crc kubenswrapper[4998]: I1203 17:13:09.785676 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-276n4"] Dec 03 17:13:10 crc kubenswrapper[4998]: I1203 17:13:10.710704 4998 generic.go:334] "Generic (PLEG): container finished" podID="26568f59-e843-42cd-b87d-3da76a9d622e" containerID="d9852c7ac707d915cd27694fca3c394d14eb1eea1d29b6e98db1b9dee49b3dd8" exitCode=0 Dec 03 17:13:10 crc kubenswrapper[4998]: I1203 17:13:10.710788 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-276n4" event={"ID":"26568f59-e843-42cd-b87d-3da76a9d622e","Type":"ContainerDied","Data":"d9852c7ac707d915cd27694fca3c394d14eb1eea1d29b6e98db1b9dee49b3dd8"} Dec 03 17:13:10 crc kubenswrapper[4998]: I1203 17:13:10.711241 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-276n4" event={"ID":"26568f59-e843-42cd-b87d-3da76a9d622e","Type":"ContainerStarted","Data":"87c287e183c4e769b24448941b3415bc96c858386cd149c2a78423b62f0ee85f"} Dec 03 17:13:10 crc kubenswrapper[4998]: I1203 17:13:10.713637 4998 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 17:13:11 crc kubenswrapper[4998]: I1203 17:13:11.730732 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-276n4" event={"ID":"26568f59-e843-42cd-b87d-3da76a9d622e","Type":"ContainerStarted","Data":"5c382418cd117cfbb9dffafdc02c664699f910089593923cadbcec38eca24195"} Dec 03 17:13:14 crc kubenswrapper[4998]: I1203 17:13:14.037568 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xf6s8"] Dec 03 17:13:14 crc kubenswrapper[4998]: I1203 17:13:14.040235 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xf6s8" Dec 03 17:13:14 crc kubenswrapper[4998]: I1203 17:13:14.072727 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xf6s8"] Dec 03 17:13:14 crc kubenswrapper[4998]: I1203 17:13:14.115830 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e82e83a-b9ec-49e9-b191-f38c999c8462-utilities\") pod \"community-operators-xf6s8\" (UID: \"8e82e83a-b9ec-49e9-b191-f38c999c8462\") " pod="openshift-marketplace/community-operators-xf6s8" Dec 03 17:13:14 crc kubenswrapper[4998]: I1203 17:13:14.116097 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e82e83a-b9ec-49e9-b191-f38c999c8462-catalog-content\") pod \"community-operators-xf6s8\" (UID: \"8e82e83a-b9ec-49e9-b191-f38c999c8462\") " pod="openshift-marketplace/community-operators-xf6s8" Dec 03 17:13:14 crc kubenswrapper[4998]: I1203 17:13:14.116585 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55pnq\" (UniqueName: \"kubernetes.io/projected/8e82e83a-b9ec-49e9-b191-f38c999c8462-kube-api-access-55pnq\") pod \"community-operators-xf6s8\" (UID: \"8e82e83a-b9ec-49e9-b191-f38c999c8462\") " pod="openshift-marketplace/community-operators-xf6s8" Dec 03 17:13:14 crc kubenswrapper[4998]: I1203 17:13:14.217878 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55pnq\" (UniqueName: \"kubernetes.io/projected/8e82e83a-b9ec-49e9-b191-f38c999c8462-kube-api-access-55pnq\") pod \"community-operators-xf6s8\" (UID: \"8e82e83a-b9ec-49e9-b191-f38c999c8462\") " pod="openshift-marketplace/community-operators-xf6s8" Dec 03 17:13:14 crc kubenswrapper[4998]: I1203 17:13:14.218240 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e82e83a-b9ec-49e9-b191-f38c999c8462-utilities\") pod \"community-operators-xf6s8\" (UID: \"8e82e83a-b9ec-49e9-b191-f38c999c8462\") " pod="openshift-marketplace/community-operators-xf6s8" Dec 03 17:13:14 crc kubenswrapper[4998]: I1203 17:13:14.218278 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e82e83a-b9ec-49e9-b191-f38c999c8462-catalog-content\") pod \"community-operators-xf6s8\" (UID: \"8e82e83a-b9ec-49e9-b191-f38c999c8462\") " pod="openshift-marketplace/community-operators-xf6s8" Dec 03 17:13:14 crc kubenswrapper[4998]: I1203 17:13:14.218688 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e82e83a-b9ec-49e9-b191-f38c999c8462-catalog-content\") pod \"community-operators-xf6s8\" (UID: \"8e82e83a-b9ec-49e9-b191-f38c999c8462\") " pod="openshift-marketplace/community-operators-xf6s8" Dec 03 17:13:14 crc kubenswrapper[4998]: I1203 17:13:14.218914 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e82e83a-b9ec-49e9-b191-f38c999c8462-utilities\") pod \"community-operators-xf6s8\" (UID: \"8e82e83a-b9ec-49e9-b191-f38c999c8462\") " pod="openshift-marketplace/community-operators-xf6s8" Dec 03 17:13:14 crc kubenswrapper[4998]: I1203 17:13:14.250456 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55pnq\" (UniqueName: \"kubernetes.io/projected/8e82e83a-b9ec-49e9-b191-f38c999c8462-kube-api-access-55pnq\") pod \"community-operators-xf6s8\" (UID: \"8e82e83a-b9ec-49e9-b191-f38c999c8462\") " pod="openshift-marketplace/community-operators-xf6s8" Dec 03 17:13:14 crc kubenswrapper[4998]: I1203 17:13:14.365495 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xf6s8" Dec 03 17:13:14 crc kubenswrapper[4998]: I1203 17:13:14.957808 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xf6s8"] Dec 03 17:13:15 crc kubenswrapper[4998]: I1203 17:13:15.767617 4998 generic.go:334] "Generic (PLEG): container finished" podID="26568f59-e843-42cd-b87d-3da76a9d622e" containerID="5c382418cd117cfbb9dffafdc02c664699f910089593923cadbcec38eca24195" exitCode=0 Dec 03 17:13:15 crc kubenswrapper[4998]: I1203 17:13:15.767710 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-276n4" event={"ID":"26568f59-e843-42cd-b87d-3da76a9d622e","Type":"ContainerDied","Data":"5c382418cd117cfbb9dffafdc02c664699f910089593923cadbcec38eca24195"} Dec 03 17:13:15 crc kubenswrapper[4998]: I1203 17:13:15.770405 4998 generic.go:334] "Generic (PLEG): container finished" podID="8e82e83a-b9ec-49e9-b191-f38c999c8462" containerID="6d37720622c0ac42cafb01031da25b178993d3e5a504edb656a67436e71e64d6" exitCode=0 Dec 03 17:13:15 crc kubenswrapper[4998]: I1203 17:13:15.770642 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xf6s8" event={"ID":"8e82e83a-b9ec-49e9-b191-f38c999c8462","Type":"ContainerDied","Data":"6d37720622c0ac42cafb01031da25b178993d3e5a504edb656a67436e71e64d6"} Dec 03 17:13:15 crc kubenswrapper[4998]: I1203 17:13:15.772854 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xf6s8" event={"ID":"8e82e83a-b9ec-49e9-b191-f38c999c8462","Type":"ContainerStarted","Data":"1ffa6d1b2567b610084ceeaa954e5f3c11057a70a8f54e8ef5a0732fd78cf7e0"} Dec 03 17:13:17 crc kubenswrapper[4998]: I1203 17:13:17.793844 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-276n4" event={"ID":"26568f59-e843-42cd-b87d-3da76a9d622e","Type":"ContainerStarted","Data":"d7c60711e6baa20508897887238f9adcb6513ff75c7abac8cca54b5acdd202f7"} Dec 03 17:13:17 crc kubenswrapper[4998]: I1203 17:13:17.799514 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xf6s8" event={"ID":"8e82e83a-b9ec-49e9-b191-f38c999c8462","Type":"ContainerStarted","Data":"d4b551c2717ec060e8fba1d6f63443afb0560e501a8967d065dfa76e463d1d3a"} Dec 03 17:13:17 crc kubenswrapper[4998]: I1203 17:13:17.816809 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-276n4" podStartSLOduration=4.113275455 podStartE2EDuration="9.816790546s" podCreationTimestamp="2025-12-03 17:13:08 +0000 UTC" firstStartedPulling="2025-12-03 17:13:10.713388097 +0000 UTC m=+4169.325088320" lastFinishedPulling="2025-12-03 17:13:16.416903188 +0000 UTC m=+4175.028603411" observedRunningTime="2025-12-03 17:13:17.815455413 +0000 UTC m=+4176.427155636" watchObservedRunningTime="2025-12-03 17:13:17.816790546 +0000 UTC m=+4176.428490769" Dec 03 17:13:19 crc kubenswrapper[4998]: I1203 17:13:19.297621 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-276n4" Dec 03 17:13:19 crc kubenswrapper[4998]: I1203 17:13:19.297935 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-276n4" Dec 03 17:13:20 crc kubenswrapper[4998]: I1203 17:13:20.542017 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-276n4" podUID="26568f59-e843-42cd-b87d-3da76a9d622e" containerName="registry-server" probeResult="failure" output=< Dec 03 17:13:20 crc kubenswrapper[4998]: timeout: failed to connect service ":50051" within 1s Dec 03 17:13:20 crc kubenswrapper[4998]: > Dec 03 17:13:20 crc kubenswrapper[4998]: I1203 17:13:20.833083 4998 generic.go:334] "Generic (PLEG): container finished" podID="8e82e83a-b9ec-49e9-b191-f38c999c8462" containerID="d4b551c2717ec060e8fba1d6f63443afb0560e501a8967d065dfa76e463d1d3a" exitCode=0 Dec 03 17:13:20 crc kubenswrapper[4998]: I1203 17:13:20.833183 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xf6s8" event={"ID":"8e82e83a-b9ec-49e9-b191-f38c999c8462","Type":"ContainerDied","Data":"d4b551c2717ec060e8fba1d6f63443afb0560e501a8967d065dfa76e463d1d3a"} Dec 03 17:13:22 crc kubenswrapper[4998]: I1203 17:13:22.870546 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xf6s8" event={"ID":"8e82e83a-b9ec-49e9-b191-f38c999c8462","Type":"ContainerStarted","Data":"f8119bfbc084696c1189b343210f9c3366f379fff3cdda1a81e6b0ed532934f3"} Dec 03 17:13:22 crc kubenswrapper[4998]: I1203 17:13:22.892804 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xf6s8" podStartSLOduration=2.64138956 podStartE2EDuration="8.892781736s" podCreationTimestamp="2025-12-03 17:13:14 +0000 UTC" firstStartedPulling="2025-12-03 17:13:15.773031351 +0000 UTC m=+4174.384731574" lastFinishedPulling="2025-12-03 17:13:22.024423527 +0000 UTC m=+4180.636123750" observedRunningTime="2025-12-03 17:13:22.886448509 +0000 UTC m=+4181.498148742" watchObservedRunningTime="2025-12-03 17:13:22.892781736 +0000 UTC m=+4181.504481969" Dec 03 17:13:24 crc kubenswrapper[4998]: I1203 17:13:24.365938 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xf6s8" Dec 03 17:13:24 crc kubenswrapper[4998]: I1203 17:13:24.366209 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xf6s8" Dec 03 17:13:25 crc kubenswrapper[4998]: I1203 17:13:25.417633 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-xf6s8" podUID="8e82e83a-b9ec-49e9-b191-f38c999c8462" containerName="registry-server" probeResult="failure" output=< Dec 03 17:13:25 crc kubenswrapper[4998]: timeout: failed to connect service ":50051" within 1s Dec 03 17:13:25 crc kubenswrapper[4998]: > Dec 03 17:13:27 crc kubenswrapper[4998]: I1203 17:13:27.111543 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:13:27 crc kubenswrapper[4998]: I1203 17:13:27.111937 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:13:29 crc kubenswrapper[4998]: I1203 17:13:29.356555 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-276n4" Dec 03 17:13:29 crc kubenswrapper[4998]: I1203 17:13:29.434662 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-276n4" Dec 03 17:13:29 crc kubenswrapper[4998]: I1203 17:13:29.605470 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-276n4"] Dec 03 17:13:30 crc kubenswrapper[4998]: I1203 17:13:30.951865 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-276n4" podUID="26568f59-e843-42cd-b87d-3da76a9d622e" containerName="registry-server" containerID="cri-o://d7c60711e6baa20508897887238f9adcb6513ff75c7abac8cca54b5acdd202f7" gracePeriod=2 Dec 03 17:13:31 crc kubenswrapper[4998]: I1203 17:13:31.978201 4998 generic.go:334] "Generic (PLEG): container finished" podID="26568f59-e843-42cd-b87d-3da76a9d622e" containerID="d7c60711e6baa20508897887238f9adcb6513ff75c7abac8cca54b5acdd202f7" exitCode=0 Dec 03 17:13:31 crc kubenswrapper[4998]: I1203 17:13:31.978624 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-276n4" event={"ID":"26568f59-e843-42cd-b87d-3da76a9d622e","Type":"ContainerDied","Data":"d7c60711e6baa20508897887238f9adcb6513ff75c7abac8cca54b5acdd202f7"} Dec 03 17:13:31 crc kubenswrapper[4998]: I1203 17:13:31.978651 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-276n4" event={"ID":"26568f59-e843-42cd-b87d-3da76a9d622e","Type":"ContainerDied","Data":"87c287e183c4e769b24448941b3415bc96c858386cd149c2a78423b62f0ee85f"} Dec 03 17:13:31 crc kubenswrapper[4998]: I1203 17:13:31.978664 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87c287e183c4e769b24448941b3415bc96c858386cd149c2a78423b62f0ee85f" Dec 03 17:13:32 crc kubenswrapper[4998]: I1203 17:13:32.050194 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-276n4" Dec 03 17:13:32 crc kubenswrapper[4998]: I1203 17:13:32.247232 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfm5s\" (UniqueName: \"kubernetes.io/projected/26568f59-e843-42cd-b87d-3da76a9d622e-kube-api-access-wfm5s\") pod \"26568f59-e843-42cd-b87d-3da76a9d622e\" (UID: \"26568f59-e843-42cd-b87d-3da76a9d622e\") " Dec 03 17:13:32 crc kubenswrapper[4998]: I1203 17:13:32.247507 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26568f59-e843-42cd-b87d-3da76a9d622e-utilities\") pod \"26568f59-e843-42cd-b87d-3da76a9d622e\" (UID: \"26568f59-e843-42cd-b87d-3da76a9d622e\") " Dec 03 17:13:32 crc kubenswrapper[4998]: I1203 17:13:32.247537 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26568f59-e843-42cd-b87d-3da76a9d622e-catalog-content\") pod \"26568f59-e843-42cd-b87d-3da76a9d622e\" (UID: \"26568f59-e843-42cd-b87d-3da76a9d622e\") " Dec 03 17:13:32 crc kubenswrapper[4998]: I1203 17:13:32.248334 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26568f59-e843-42cd-b87d-3da76a9d622e-utilities" (OuterVolumeSpecName: "utilities") pod "26568f59-e843-42cd-b87d-3da76a9d622e" (UID: "26568f59-e843-42cd-b87d-3da76a9d622e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:13:32 crc kubenswrapper[4998]: I1203 17:13:32.253041 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26568f59-e843-42cd-b87d-3da76a9d622e-kube-api-access-wfm5s" (OuterVolumeSpecName: "kube-api-access-wfm5s") pod "26568f59-e843-42cd-b87d-3da76a9d622e" (UID: "26568f59-e843-42cd-b87d-3da76a9d622e"). InnerVolumeSpecName "kube-api-access-wfm5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:32 crc kubenswrapper[4998]: I1203 17:13:32.356092 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26568f59-e843-42cd-b87d-3da76a9d622e-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:32 crc kubenswrapper[4998]: I1203 17:13:32.356153 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfm5s\" (UniqueName: \"kubernetes.io/projected/26568f59-e843-42cd-b87d-3da76a9d622e-kube-api-access-wfm5s\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:32 crc kubenswrapper[4998]: I1203 17:13:32.357410 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26568f59-e843-42cd-b87d-3da76a9d622e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "26568f59-e843-42cd-b87d-3da76a9d622e" (UID: "26568f59-e843-42cd-b87d-3da76a9d622e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:13:32 crc kubenswrapper[4998]: I1203 17:13:32.458433 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26568f59-e843-42cd-b87d-3da76a9d622e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:32 crc kubenswrapper[4998]: I1203 17:13:32.990329 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-276n4" Dec 03 17:13:33 crc kubenswrapper[4998]: I1203 17:13:33.034450 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-276n4"] Dec 03 17:13:33 crc kubenswrapper[4998]: I1203 17:13:33.047846 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-276n4"] Dec 03 17:13:33 crc kubenswrapper[4998]: I1203 17:13:33.701485 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26568f59-e843-42cd-b87d-3da76a9d622e" path="/var/lib/kubelet/pods/26568f59-e843-42cd-b87d-3da76a9d622e/volumes" Dec 03 17:13:34 crc kubenswrapper[4998]: I1203 17:13:34.440596 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xf6s8" Dec 03 17:13:34 crc kubenswrapper[4998]: I1203 17:13:34.494741 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xf6s8" Dec 03 17:13:35 crc kubenswrapper[4998]: I1203 17:13:35.005065 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xf6s8"] Dec 03 17:13:36 crc kubenswrapper[4998]: I1203 17:13:36.026496 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xf6s8" podUID="8e82e83a-b9ec-49e9-b191-f38c999c8462" containerName="registry-server" containerID="cri-o://f8119bfbc084696c1189b343210f9c3366f379fff3cdda1a81e6b0ed532934f3" gracePeriod=2 Dec 03 17:13:36 crc kubenswrapper[4998]: I1203 17:13:36.548105 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xf6s8" Dec 03 17:13:36 crc kubenswrapper[4998]: I1203 17:13:36.659557 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e82e83a-b9ec-49e9-b191-f38c999c8462-catalog-content\") pod \"8e82e83a-b9ec-49e9-b191-f38c999c8462\" (UID: \"8e82e83a-b9ec-49e9-b191-f38c999c8462\") " Dec 03 17:13:36 crc kubenswrapper[4998]: I1203 17:13:36.659612 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55pnq\" (UniqueName: \"kubernetes.io/projected/8e82e83a-b9ec-49e9-b191-f38c999c8462-kube-api-access-55pnq\") pod \"8e82e83a-b9ec-49e9-b191-f38c999c8462\" (UID: \"8e82e83a-b9ec-49e9-b191-f38c999c8462\") " Dec 03 17:13:36 crc kubenswrapper[4998]: I1203 17:13:36.660038 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e82e83a-b9ec-49e9-b191-f38c999c8462-utilities\") pod \"8e82e83a-b9ec-49e9-b191-f38c999c8462\" (UID: \"8e82e83a-b9ec-49e9-b191-f38c999c8462\") " Dec 03 17:13:36 crc kubenswrapper[4998]: I1203 17:13:36.660532 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e82e83a-b9ec-49e9-b191-f38c999c8462-utilities" (OuterVolumeSpecName: "utilities") pod "8e82e83a-b9ec-49e9-b191-f38c999c8462" (UID: "8e82e83a-b9ec-49e9-b191-f38c999c8462"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:13:36 crc kubenswrapper[4998]: I1203 17:13:36.667215 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e82e83a-b9ec-49e9-b191-f38c999c8462-kube-api-access-55pnq" (OuterVolumeSpecName: "kube-api-access-55pnq") pod "8e82e83a-b9ec-49e9-b191-f38c999c8462" (UID: "8e82e83a-b9ec-49e9-b191-f38c999c8462"). InnerVolumeSpecName "kube-api-access-55pnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:36 crc kubenswrapper[4998]: I1203 17:13:36.732743 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e82e83a-b9ec-49e9-b191-f38c999c8462-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e82e83a-b9ec-49e9-b191-f38c999c8462" (UID: "8e82e83a-b9ec-49e9-b191-f38c999c8462"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:13:36 crc kubenswrapper[4998]: I1203 17:13:36.762490 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e82e83a-b9ec-49e9-b191-f38c999c8462-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:36 crc kubenswrapper[4998]: I1203 17:13:36.762738 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e82e83a-b9ec-49e9-b191-f38c999c8462-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:36 crc kubenswrapper[4998]: I1203 17:13:36.762828 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55pnq\" (UniqueName: \"kubernetes.io/projected/8e82e83a-b9ec-49e9-b191-f38c999c8462-kube-api-access-55pnq\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:37 crc kubenswrapper[4998]: I1203 17:13:37.039858 4998 generic.go:334] "Generic (PLEG): container finished" podID="8e82e83a-b9ec-49e9-b191-f38c999c8462" containerID="f8119bfbc084696c1189b343210f9c3366f379fff3cdda1a81e6b0ed532934f3" exitCode=0 Dec 03 17:13:37 crc kubenswrapper[4998]: I1203 17:13:37.039960 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xf6s8" Dec 03 17:13:37 crc kubenswrapper[4998]: I1203 17:13:37.039937 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xf6s8" event={"ID":"8e82e83a-b9ec-49e9-b191-f38c999c8462","Type":"ContainerDied","Data":"f8119bfbc084696c1189b343210f9c3366f379fff3cdda1a81e6b0ed532934f3"} Dec 03 17:13:37 crc kubenswrapper[4998]: I1203 17:13:37.040322 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xf6s8" event={"ID":"8e82e83a-b9ec-49e9-b191-f38c999c8462","Type":"ContainerDied","Data":"1ffa6d1b2567b610084ceeaa954e5f3c11057a70a8f54e8ef5a0732fd78cf7e0"} Dec 03 17:13:37 crc kubenswrapper[4998]: I1203 17:13:37.040357 4998 scope.go:117] "RemoveContainer" containerID="f8119bfbc084696c1189b343210f9c3366f379fff3cdda1a81e6b0ed532934f3" Dec 03 17:13:37 crc kubenswrapper[4998]: I1203 17:13:37.075457 4998 scope.go:117] "RemoveContainer" containerID="d4b551c2717ec060e8fba1d6f63443afb0560e501a8967d065dfa76e463d1d3a" Dec 03 17:13:37 crc kubenswrapper[4998]: I1203 17:13:37.082343 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xf6s8"] Dec 03 17:13:37 crc kubenswrapper[4998]: I1203 17:13:37.095642 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xf6s8"] Dec 03 17:13:37 crc kubenswrapper[4998]: I1203 17:13:37.104630 4998 scope.go:117] "RemoveContainer" containerID="6d37720622c0ac42cafb01031da25b178993d3e5a504edb656a67436e71e64d6" Dec 03 17:13:37 crc kubenswrapper[4998]: I1203 17:13:37.177369 4998 scope.go:117] "RemoveContainer" containerID="f8119bfbc084696c1189b343210f9c3366f379fff3cdda1a81e6b0ed532934f3" Dec 03 17:13:37 crc kubenswrapper[4998]: E1203 17:13:37.178060 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8119bfbc084696c1189b343210f9c3366f379fff3cdda1a81e6b0ed532934f3\": container with ID starting with f8119bfbc084696c1189b343210f9c3366f379fff3cdda1a81e6b0ed532934f3 not found: ID does not exist" containerID="f8119bfbc084696c1189b343210f9c3366f379fff3cdda1a81e6b0ed532934f3" Dec 03 17:13:37 crc kubenswrapper[4998]: I1203 17:13:37.178118 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8119bfbc084696c1189b343210f9c3366f379fff3cdda1a81e6b0ed532934f3"} err="failed to get container status \"f8119bfbc084696c1189b343210f9c3366f379fff3cdda1a81e6b0ed532934f3\": rpc error: code = NotFound desc = could not find container \"f8119bfbc084696c1189b343210f9c3366f379fff3cdda1a81e6b0ed532934f3\": container with ID starting with f8119bfbc084696c1189b343210f9c3366f379fff3cdda1a81e6b0ed532934f3 not found: ID does not exist" Dec 03 17:13:37 crc kubenswrapper[4998]: I1203 17:13:37.178187 4998 scope.go:117] "RemoveContainer" containerID="d4b551c2717ec060e8fba1d6f63443afb0560e501a8967d065dfa76e463d1d3a" Dec 03 17:13:37 crc kubenswrapper[4998]: E1203 17:13:37.178826 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4b551c2717ec060e8fba1d6f63443afb0560e501a8967d065dfa76e463d1d3a\": container with ID starting with d4b551c2717ec060e8fba1d6f63443afb0560e501a8967d065dfa76e463d1d3a not found: ID does not exist" containerID="d4b551c2717ec060e8fba1d6f63443afb0560e501a8967d065dfa76e463d1d3a" Dec 03 17:13:37 crc kubenswrapper[4998]: I1203 17:13:37.178909 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4b551c2717ec060e8fba1d6f63443afb0560e501a8967d065dfa76e463d1d3a"} err="failed to get container status \"d4b551c2717ec060e8fba1d6f63443afb0560e501a8967d065dfa76e463d1d3a\": rpc error: code = NotFound desc = could not find container \"d4b551c2717ec060e8fba1d6f63443afb0560e501a8967d065dfa76e463d1d3a\": container with ID starting with d4b551c2717ec060e8fba1d6f63443afb0560e501a8967d065dfa76e463d1d3a not found: ID does not exist" Dec 03 17:13:37 crc kubenswrapper[4998]: I1203 17:13:37.178957 4998 scope.go:117] "RemoveContainer" containerID="6d37720622c0ac42cafb01031da25b178993d3e5a504edb656a67436e71e64d6" Dec 03 17:13:37 crc kubenswrapper[4998]: E1203 17:13:37.179609 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d37720622c0ac42cafb01031da25b178993d3e5a504edb656a67436e71e64d6\": container with ID starting with 6d37720622c0ac42cafb01031da25b178993d3e5a504edb656a67436e71e64d6 not found: ID does not exist" containerID="6d37720622c0ac42cafb01031da25b178993d3e5a504edb656a67436e71e64d6" Dec 03 17:13:37 crc kubenswrapper[4998]: I1203 17:13:37.179692 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d37720622c0ac42cafb01031da25b178993d3e5a504edb656a67436e71e64d6"} err="failed to get container status \"6d37720622c0ac42cafb01031da25b178993d3e5a504edb656a67436e71e64d6\": rpc error: code = NotFound desc = could not find container \"6d37720622c0ac42cafb01031da25b178993d3e5a504edb656a67436e71e64d6\": container with ID starting with 6d37720622c0ac42cafb01031da25b178993d3e5a504edb656a67436e71e64d6 not found: ID does not exist" Dec 03 17:13:37 crc kubenswrapper[4998]: I1203 17:13:37.696908 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e82e83a-b9ec-49e9-b191-f38c999c8462" path="/var/lib/kubelet/pods/8e82e83a-b9ec-49e9-b191-f38c999c8462/volumes" Dec 03 17:13:57 crc kubenswrapper[4998]: I1203 17:13:57.111361 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:13:57 crc kubenswrapper[4998]: I1203 17:13:57.112144 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:14:27 crc kubenswrapper[4998]: I1203 17:14:27.111358 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:14:27 crc kubenswrapper[4998]: I1203 17:14:27.112101 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:14:27 crc kubenswrapper[4998]: I1203 17:14:27.112173 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 17:14:27 crc kubenswrapper[4998]: I1203 17:14:27.113477 4998 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85"} pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:14:27 crc kubenswrapper[4998]: I1203 17:14:27.113581 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" containerID="cri-o://3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" gracePeriod=600 Dec 03 17:14:27 crc kubenswrapper[4998]: E1203 17:14:27.285968 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:14:27 crc kubenswrapper[4998]: I1203 17:14:27.611594 4998 generic.go:334] "Generic (PLEG): container finished" podID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" exitCode=0 Dec 03 17:14:27 crc kubenswrapper[4998]: I1203 17:14:27.611673 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerDied","Data":"3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85"} Dec 03 17:14:27 crc kubenswrapper[4998]: I1203 17:14:27.611739 4998 scope.go:117] "RemoveContainer" containerID="9ac7668007b277a203e4765518b470f255fddd0cb75042129bf6f6ab95cc4cd4" Dec 03 17:14:27 crc kubenswrapper[4998]: I1203 17:14:27.613038 4998 scope.go:117] "RemoveContainer" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" Dec 03 17:14:27 crc kubenswrapper[4998]: E1203 17:14:27.613721 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:14:40 crc kubenswrapper[4998]: I1203 17:14:40.678373 4998 scope.go:117] "RemoveContainer" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" Dec 03 17:14:40 crc kubenswrapper[4998]: E1203 17:14:40.679601 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:14:53 crc kubenswrapper[4998]: I1203 17:14:53.884791 4998 scope.go:117] "RemoveContainer" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" Dec 03 17:14:53 crc kubenswrapper[4998]: E1203 17:14:53.897928 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:15:00 crc kubenswrapper[4998]: I1203 17:15:00.188142 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413035-7lxk4"] Dec 03 17:15:00 crc kubenswrapper[4998]: E1203 17:15:00.190143 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e82e83a-b9ec-49e9-b191-f38c999c8462" containerName="extract-content" Dec 03 17:15:00 crc kubenswrapper[4998]: I1203 17:15:00.190220 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e82e83a-b9ec-49e9-b191-f38c999c8462" containerName="extract-content" Dec 03 17:15:00 crc kubenswrapper[4998]: E1203 17:15:00.190290 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e82e83a-b9ec-49e9-b191-f38c999c8462" containerName="registry-server" Dec 03 17:15:00 crc kubenswrapper[4998]: I1203 17:15:00.190346 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e82e83a-b9ec-49e9-b191-f38c999c8462" containerName="registry-server" Dec 03 17:15:00 crc kubenswrapper[4998]: E1203 17:15:00.190404 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e82e83a-b9ec-49e9-b191-f38c999c8462" containerName="extract-utilities" Dec 03 17:15:00 crc kubenswrapper[4998]: I1203 17:15:00.190460 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e82e83a-b9ec-49e9-b191-f38c999c8462" containerName="extract-utilities" Dec 03 17:15:00 crc kubenswrapper[4998]: E1203 17:15:00.190538 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26568f59-e843-42cd-b87d-3da76a9d622e" containerName="extract-content" Dec 03 17:15:00 crc kubenswrapper[4998]: I1203 17:15:00.190595 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="26568f59-e843-42cd-b87d-3da76a9d622e" containerName="extract-content" Dec 03 17:15:00 crc kubenswrapper[4998]: E1203 17:15:00.190658 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26568f59-e843-42cd-b87d-3da76a9d622e" containerName="registry-server" Dec 03 17:15:00 crc kubenswrapper[4998]: I1203 17:15:00.190710 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="26568f59-e843-42cd-b87d-3da76a9d622e" containerName="registry-server" Dec 03 17:15:00 crc kubenswrapper[4998]: E1203 17:15:00.190795 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26568f59-e843-42cd-b87d-3da76a9d622e" containerName="extract-utilities" Dec 03 17:15:00 crc kubenswrapper[4998]: I1203 17:15:00.190858 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="26568f59-e843-42cd-b87d-3da76a9d622e" containerName="extract-utilities" Dec 03 17:15:00 crc kubenswrapper[4998]: I1203 17:15:00.191101 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="26568f59-e843-42cd-b87d-3da76a9d622e" containerName="registry-server" Dec 03 17:15:00 crc kubenswrapper[4998]: I1203 17:15:00.191199 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e82e83a-b9ec-49e9-b191-f38c999c8462" containerName="registry-server" Dec 03 17:15:00 crc kubenswrapper[4998]: I1203 17:15:00.192164 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-7lxk4" Dec 03 17:15:00 crc kubenswrapper[4998]: I1203 17:15:00.194948 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 17:15:00 crc kubenswrapper[4998]: I1203 17:15:00.195464 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 17:15:00 crc kubenswrapper[4998]: I1203 17:15:00.205085 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413035-7lxk4"] Dec 03 17:15:00 crc kubenswrapper[4998]: I1203 17:15:00.380257 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f32583b1-9dc6-47ef-894b-0f70b4f757e6-secret-volume\") pod \"collect-profiles-29413035-7lxk4\" (UID: \"f32583b1-9dc6-47ef-894b-0f70b4f757e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-7lxk4" Dec 03 17:15:00 crc kubenswrapper[4998]: I1203 17:15:00.380717 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f32583b1-9dc6-47ef-894b-0f70b4f757e6-config-volume\") pod \"collect-profiles-29413035-7lxk4\" (UID: \"f32583b1-9dc6-47ef-894b-0f70b4f757e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-7lxk4" Dec 03 17:15:00 crc kubenswrapper[4998]: I1203 17:15:00.380962 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6sg8\" (UniqueName: \"kubernetes.io/projected/f32583b1-9dc6-47ef-894b-0f70b4f757e6-kube-api-access-h6sg8\") pod \"collect-profiles-29413035-7lxk4\" (UID: \"f32583b1-9dc6-47ef-894b-0f70b4f757e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-7lxk4" Dec 03 17:15:00 crc kubenswrapper[4998]: I1203 17:15:00.483234 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f32583b1-9dc6-47ef-894b-0f70b4f757e6-secret-volume\") pod \"collect-profiles-29413035-7lxk4\" (UID: \"f32583b1-9dc6-47ef-894b-0f70b4f757e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-7lxk4" Dec 03 17:15:00 crc kubenswrapper[4998]: I1203 17:15:00.483525 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f32583b1-9dc6-47ef-894b-0f70b4f757e6-config-volume\") pod \"collect-profiles-29413035-7lxk4\" (UID: \"f32583b1-9dc6-47ef-894b-0f70b4f757e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-7lxk4" Dec 03 17:15:00 crc kubenswrapper[4998]: I1203 17:15:00.483648 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6sg8\" (UniqueName: \"kubernetes.io/projected/f32583b1-9dc6-47ef-894b-0f70b4f757e6-kube-api-access-h6sg8\") pod \"collect-profiles-29413035-7lxk4\" (UID: \"f32583b1-9dc6-47ef-894b-0f70b4f757e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-7lxk4" Dec 03 17:15:00 crc kubenswrapper[4998]: I1203 17:15:00.485321 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f32583b1-9dc6-47ef-894b-0f70b4f757e6-config-volume\") pod \"collect-profiles-29413035-7lxk4\" (UID: \"f32583b1-9dc6-47ef-894b-0f70b4f757e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-7lxk4" Dec 03 17:15:00 crc kubenswrapper[4998]: I1203 17:15:00.500398 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f32583b1-9dc6-47ef-894b-0f70b4f757e6-secret-volume\") pod \"collect-profiles-29413035-7lxk4\" (UID: \"f32583b1-9dc6-47ef-894b-0f70b4f757e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-7lxk4" Dec 03 17:15:00 crc kubenswrapper[4998]: I1203 17:15:00.517519 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6sg8\" (UniqueName: \"kubernetes.io/projected/f32583b1-9dc6-47ef-894b-0f70b4f757e6-kube-api-access-h6sg8\") pod \"collect-profiles-29413035-7lxk4\" (UID: \"f32583b1-9dc6-47ef-894b-0f70b4f757e6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-7lxk4" Dec 03 17:15:00 crc kubenswrapper[4998]: I1203 17:15:00.522373 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-7lxk4" Dec 03 17:15:01 crc kubenswrapper[4998]: I1203 17:15:01.021728 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413035-7lxk4"] Dec 03 17:15:02 crc kubenswrapper[4998]: I1203 17:15:02.024098 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-7lxk4" event={"ID":"f32583b1-9dc6-47ef-894b-0f70b4f757e6","Type":"ContainerStarted","Data":"ca837beae19878170d12aa8c25837f1675cd6b8daf06da08719388fbae553cc5"} Dec 03 17:15:02 crc kubenswrapper[4998]: I1203 17:15:02.024349 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-7lxk4" event={"ID":"f32583b1-9dc6-47ef-894b-0f70b4f757e6","Type":"ContainerStarted","Data":"ee971ae4b4c07a2bef442c8688873bf163323a74a4f546c735e6e00169801725"} Dec 03 17:15:02 crc kubenswrapper[4998]: I1203 17:15:02.054623 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-7lxk4" podStartSLOduration=2.054599331 podStartE2EDuration="2.054599331s" podCreationTimestamp="2025-12-03 17:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:02.04610322 +0000 UTC m=+4280.657803453" watchObservedRunningTime="2025-12-03 17:15:02.054599331 +0000 UTC m=+4280.666299574" Dec 03 17:15:03 crc kubenswrapper[4998]: I1203 17:15:03.038178 4998 generic.go:334] "Generic (PLEG): container finished" podID="f32583b1-9dc6-47ef-894b-0f70b4f757e6" containerID="ca837beae19878170d12aa8c25837f1675cd6b8daf06da08719388fbae553cc5" exitCode=0 Dec 03 17:15:03 crc kubenswrapper[4998]: I1203 17:15:03.038238 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-7lxk4" event={"ID":"f32583b1-9dc6-47ef-894b-0f70b4f757e6","Type":"ContainerDied","Data":"ca837beae19878170d12aa8c25837f1675cd6b8daf06da08719388fbae553cc5"} Dec 03 17:15:04 crc kubenswrapper[4998]: I1203 17:15:04.419534 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-7lxk4" Dec 03 17:15:04 crc kubenswrapper[4998]: I1203 17:15:04.479667 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f32583b1-9dc6-47ef-894b-0f70b4f757e6-config-volume\") pod \"f32583b1-9dc6-47ef-894b-0f70b4f757e6\" (UID: \"f32583b1-9dc6-47ef-894b-0f70b4f757e6\") " Dec 03 17:15:04 crc kubenswrapper[4998]: I1203 17:15:04.479735 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f32583b1-9dc6-47ef-894b-0f70b4f757e6-secret-volume\") pod \"f32583b1-9dc6-47ef-894b-0f70b4f757e6\" (UID: \"f32583b1-9dc6-47ef-894b-0f70b4f757e6\") " Dec 03 17:15:04 crc kubenswrapper[4998]: I1203 17:15:04.480209 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6sg8\" (UniqueName: \"kubernetes.io/projected/f32583b1-9dc6-47ef-894b-0f70b4f757e6-kube-api-access-h6sg8\") pod \"f32583b1-9dc6-47ef-894b-0f70b4f757e6\" (UID: \"f32583b1-9dc6-47ef-894b-0f70b4f757e6\") " Dec 03 17:15:04 crc kubenswrapper[4998]: I1203 17:15:04.484289 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f32583b1-9dc6-47ef-894b-0f70b4f757e6-config-volume" (OuterVolumeSpecName: "config-volume") pod "f32583b1-9dc6-47ef-894b-0f70b4f757e6" (UID: "f32583b1-9dc6-47ef-894b-0f70b4f757e6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:15:04 crc kubenswrapper[4998]: I1203 17:15:04.493020 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f32583b1-9dc6-47ef-894b-0f70b4f757e6-kube-api-access-h6sg8" (OuterVolumeSpecName: "kube-api-access-h6sg8") pod "f32583b1-9dc6-47ef-894b-0f70b4f757e6" (UID: "f32583b1-9dc6-47ef-894b-0f70b4f757e6"). InnerVolumeSpecName "kube-api-access-h6sg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:15:04 crc kubenswrapper[4998]: I1203 17:15:04.493134 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f32583b1-9dc6-47ef-894b-0f70b4f757e6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f32583b1-9dc6-47ef-894b-0f70b4f757e6" (UID: "f32583b1-9dc6-47ef-894b-0f70b4f757e6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:15:04 crc kubenswrapper[4998]: I1203 17:15:04.582487 4998 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f32583b1-9dc6-47ef-894b-0f70b4f757e6-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 17:15:04 crc kubenswrapper[4998]: I1203 17:15:04.582802 4998 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f32583b1-9dc6-47ef-894b-0f70b4f757e6-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 17:15:04 crc kubenswrapper[4998]: I1203 17:15:04.582881 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6sg8\" (UniqueName: \"kubernetes.io/projected/f32583b1-9dc6-47ef-894b-0f70b4f757e6-kube-api-access-h6sg8\") on node \"crc\" DevicePath \"\"" Dec 03 17:15:04 crc kubenswrapper[4998]: I1203 17:15:04.797067 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412990-zcxf2"] Dec 03 17:15:04 crc kubenswrapper[4998]: I1203 17:15:04.811418 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412990-zcxf2"] Dec 03 17:15:05 crc kubenswrapper[4998]: I1203 17:15:05.073577 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-7lxk4" event={"ID":"f32583b1-9dc6-47ef-894b-0f70b4f757e6","Type":"ContainerDied","Data":"ee971ae4b4c07a2bef442c8688873bf163323a74a4f546c735e6e00169801725"} Dec 03 17:15:05 crc kubenswrapper[4998]: I1203 17:15:05.073931 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee971ae4b4c07a2bef442c8688873bf163323a74a4f546c735e6e00169801725" Dec 03 17:15:05 crc kubenswrapper[4998]: I1203 17:15:05.073697 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-7lxk4" Dec 03 17:15:05 crc kubenswrapper[4998]: I1203 17:15:05.698749 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0" path="/var/lib/kubelet/pods/bd7e24f1-cbc1-4962-8b5b-f1cc1afa2dc0/volumes" Dec 03 17:15:07 crc kubenswrapper[4998]: I1203 17:15:07.678504 4998 scope.go:117] "RemoveContainer" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" Dec 03 17:15:07 crc kubenswrapper[4998]: E1203 17:15:07.679141 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:15:18 crc kubenswrapper[4998]: I1203 17:15:18.678801 4998 scope.go:117] "RemoveContainer" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" Dec 03 17:15:18 crc kubenswrapper[4998]: E1203 17:15:18.679578 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:15:29 crc kubenswrapper[4998]: I1203 17:15:29.677948 4998 scope.go:117] "RemoveContainer" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" Dec 03 17:15:29 crc kubenswrapper[4998]: E1203 17:15:29.679067 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:15:44 crc kubenswrapper[4998]: I1203 17:15:44.678109 4998 scope.go:117] "RemoveContainer" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" Dec 03 17:15:44 crc kubenswrapper[4998]: E1203 17:15:44.678874 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:15:58 crc kubenswrapper[4998]: I1203 17:15:58.677447 4998 scope.go:117] "RemoveContainer" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" Dec 03 17:15:58 crc kubenswrapper[4998]: E1203 17:15:58.678312 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:16:01 crc kubenswrapper[4998]: I1203 17:16:01.814476 4998 scope.go:117] "RemoveContainer" containerID="b0f6a6a2769945bfb5e176d55d2f18bbd755e2fc39a356d4b06808ce4c40b542" Dec 03 17:16:09 crc kubenswrapper[4998]: I1203 17:16:09.678935 4998 scope.go:117] "RemoveContainer" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" Dec 03 17:16:09 crc kubenswrapper[4998]: E1203 17:16:09.680310 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:16:22 crc kubenswrapper[4998]: I1203 17:16:22.678443 4998 scope.go:117] "RemoveContainer" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" Dec 03 17:16:22 crc kubenswrapper[4998]: E1203 17:16:22.680333 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:16:35 crc kubenswrapper[4998]: I1203 17:16:35.678196 4998 scope.go:117] "RemoveContainer" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" Dec 03 17:16:35 crc kubenswrapper[4998]: E1203 17:16:35.679576 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:16:46 crc kubenswrapper[4998]: I1203 17:16:46.677497 4998 scope.go:117] "RemoveContainer" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" Dec 03 17:16:46 crc kubenswrapper[4998]: E1203 17:16:46.678293 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:16:59 crc kubenswrapper[4998]: I1203 17:16:59.154307 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4cmzb"] Dec 03 17:16:59 crc kubenswrapper[4998]: E1203 17:16:59.155370 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f32583b1-9dc6-47ef-894b-0f70b4f757e6" containerName="collect-profiles" Dec 03 17:16:59 crc kubenswrapper[4998]: I1203 17:16:59.155387 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="f32583b1-9dc6-47ef-894b-0f70b4f757e6" containerName="collect-profiles" Dec 03 17:16:59 crc kubenswrapper[4998]: I1203 17:16:59.155638 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="f32583b1-9dc6-47ef-894b-0f70b4f757e6" containerName="collect-profiles" Dec 03 17:16:59 crc kubenswrapper[4998]: I1203 17:16:59.157508 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4cmzb" Dec 03 17:16:59 crc kubenswrapper[4998]: I1203 17:16:59.169964 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4cmzb"] Dec 03 17:16:59 crc kubenswrapper[4998]: I1203 17:16:59.269402 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rj6b\" (UniqueName: \"kubernetes.io/projected/c7ff8f3a-debc-42a6-b564-4fc470189f24-kube-api-access-9rj6b\") pod \"certified-operators-4cmzb\" (UID: \"c7ff8f3a-debc-42a6-b564-4fc470189f24\") " pod="openshift-marketplace/certified-operators-4cmzb" Dec 03 17:16:59 crc kubenswrapper[4998]: I1203 17:16:59.269540 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7ff8f3a-debc-42a6-b564-4fc470189f24-catalog-content\") pod \"certified-operators-4cmzb\" (UID: \"c7ff8f3a-debc-42a6-b564-4fc470189f24\") " pod="openshift-marketplace/certified-operators-4cmzb" Dec 03 17:16:59 crc kubenswrapper[4998]: I1203 17:16:59.269564 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7ff8f3a-debc-42a6-b564-4fc470189f24-utilities\") pod \"certified-operators-4cmzb\" (UID: \"c7ff8f3a-debc-42a6-b564-4fc470189f24\") " pod="openshift-marketplace/certified-operators-4cmzb" Dec 03 17:16:59 crc kubenswrapper[4998]: I1203 17:16:59.371387 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7ff8f3a-debc-42a6-b564-4fc470189f24-catalog-content\") pod \"certified-operators-4cmzb\" (UID: \"c7ff8f3a-debc-42a6-b564-4fc470189f24\") " pod="openshift-marketplace/certified-operators-4cmzb" Dec 03 17:16:59 crc kubenswrapper[4998]: I1203 17:16:59.371441 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7ff8f3a-debc-42a6-b564-4fc470189f24-utilities\") pod \"certified-operators-4cmzb\" (UID: \"c7ff8f3a-debc-42a6-b564-4fc470189f24\") " pod="openshift-marketplace/certified-operators-4cmzb" Dec 03 17:16:59 crc kubenswrapper[4998]: I1203 17:16:59.371616 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rj6b\" (UniqueName: \"kubernetes.io/projected/c7ff8f3a-debc-42a6-b564-4fc470189f24-kube-api-access-9rj6b\") pod \"certified-operators-4cmzb\" (UID: \"c7ff8f3a-debc-42a6-b564-4fc470189f24\") " pod="openshift-marketplace/certified-operators-4cmzb" Dec 03 17:16:59 crc kubenswrapper[4998]: I1203 17:16:59.372213 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7ff8f3a-debc-42a6-b564-4fc470189f24-catalog-content\") pod \"certified-operators-4cmzb\" (UID: \"c7ff8f3a-debc-42a6-b564-4fc470189f24\") " pod="openshift-marketplace/certified-operators-4cmzb" Dec 03 17:16:59 crc kubenswrapper[4998]: I1203 17:16:59.372238 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7ff8f3a-debc-42a6-b564-4fc470189f24-utilities\") pod \"certified-operators-4cmzb\" (UID: \"c7ff8f3a-debc-42a6-b564-4fc470189f24\") " pod="openshift-marketplace/certified-operators-4cmzb" Dec 03 17:16:59 crc kubenswrapper[4998]: I1203 17:16:59.401503 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rj6b\" (UniqueName: \"kubernetes.io/projected/c7ff8f3a-debc-42a6-b564-4fc470189f24-kube-api-access-9rj6b\") pod \"certified-operators-4cmzb\" (UID: \"c7ff8f3a-debc-42a6-b564-4fc470189f24\") " pod="openshift-marketplace/certified-operators-4cmzb" Dec 03 17:16:59 crc kubenswrapper[4998]: I1203 17:16:59.498464 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4cmzb" Dec 03 17:16:59 crc kubenswrapper[4998]: I1203 17:16:59.687683 4998 scope.go:117] "RemoveContainer" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" Dec 03 17:16:59 crc kubenswrapper[4998]: E1203 17:16:59.688339 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:17:00 crc kubenswrapper[4998]: I1203 17:17:00.064231 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4cmzb"] Dec 03 17:17:00 crc kubenswrapper[4998]: I1203 17:17:00.269430 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4cmzb" event={"ID":"c7ff8f3a-debc-42a6-b564-4fc470189f24","Type":"ContainerStarted","Data":"3ee0364cffef6dcf72b5e9a22d7fa75eb0b809decdd9189e9183e1707141e8da"} Dec 03 17:17:01 crc kubenswrapper[4998]: I1203 17:17:01.280808 4998 generic.go:334] "Generic (PLEG): container finished" podID="c7ff8f3a-debc-42a6-b564-4fc470189f24" containerID="1da7a3c5628448b6e2a55db1b875092e2488adf9f80013ddc4347fd7e0847764" exitCode=0 Dec 03 17:17:01 crc kubenswrapper[4998]: I1203 17:17:01.280978 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4cmzb" event={"ID":"c7ff8f3a-debc-42a6-b564-4fc470189f24","Type":"ContainerDied","Data":"1da7a3c5628448b6e2a55db1b875092e2488adf9f80013ddc4347fd7e0847764"} Dec 03 17:17:03 crc kubenswrapper[4998]: I1203 17:17:03.302577 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4cmzb" event={"ID":"c7ff8f3a-debc-42a6-b564-4fc470189f24","Type":"ContainerStarted","Data":"1082bebec6add1d63556b7af1d17f550cbd8ad09d6cf7f53bf5cda8f9fe25654"} Dec 03 17:17:04 crc kubenswrapper[4998]: I1203 17:17:04.316105 4998 generic.go:334] "Generic (PLEG): container finished" podID="c7ff8f3a-debc-42a6-b564-4fc470189f24" containerID="1082bebec6add1d63556b7af1d17f550cbd8ad09d6cf7f53bf5cda8f9fe25654" exitCode=0 Dec 03 17:17:04 crc kubenswrapper[4998]: I1203 17:17:04.316183 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4cmzb" event={"ID":"c7ff8f3a-debc-42a6-b564-4fc470189f24","Type":"ContainerDied","Data":"1082bebec6add1d63556b7af1d17f550cbd8ad09d6cf7f53bf5cda8f9fe25654"} Dec 03 17:17:06 crc kubenswrapper[4998]: I1203 17:17:06.338065 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4cmzb" event={"ID":"c7ff8f3a-debc-42a6-b564-4fc470189f24","Type":"ContainerStarted","Data":"6f9b1da1e59db40a0f75ec91965a14914e6d33d0752398dd61625a856ae20a3b"} Dec 03 17:17:06 crc kubenswrapper[4998]: I1203 17:17:06.360943 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4cmzb" podStartSLOduration=3.852599923 podStartE2EDuration="7.360921811s" podCreationTimestamp="2025-12-03 17:16:59 +0000 UTC" firstStartedPulling="2025-12-03 17:17:01.303455321 +0000 UTC m=+4399.915155544" lastFinishedPulling="2025-12-03 17:17:04.811777209 +0000 UTC m=+4403.423477432" observedRunningTime="2025-12-03 17:17:06.35401536 +0000 UTC m=+4404.965715593" watchObservedRunningTime="2025-12-03 17:17:06.360921811 +0000 UTC m=+4404.972622034" Dec 03 17:17:09 crc kubenswrapper[4998]: I1203 17:17:09.499173 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4cmzb" Dec 03 17:17:09 crc kubenswrapper[4998]: I1203 17:17:09.500090 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4cmzb" Dec 03 17:17:09 crc kubenswrapper[4998]: I1203 17:17:09.550454 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4cmzb" Dec 03 17:17:14 crc kubenswrapper[4998]: I1203 17:17:14.678488 4998 scope.go:117] "RemoveContainer" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" Dec 03 17:17:14 crc kubenswrapper[4998]: E1203 17:17:14.679403 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:17:19 crc kubenswrapper[4998]: I1203 17:17:19.548183 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4cmzb" Dec 03 17:17:19 crc kubenswrapper[4998]: I1203 17:17:19.628243 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4cmzb"] Dec 03 17:17:20 crc kubenswrapper[4998]: I1203 17:17:20.480142 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4cmzb" podUID="c7ff8f3a-debc-42a6-b564-4fc470189f24" containerName="registry-server" containerID="cri-o://6f9b1da1e59db40a0f75ec91965a14914e6d33d0752398dd61625a856ae20a3b" gracePeriod=2 Dec 03 17:17:20 crc kubenswrapper[4998]: I1203 17:17:20.965476 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4cmzb" Dec 03 17:17:21 crc kubenswrapper[4998]: I1203 17:17:21.116356 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7ff8f3a-debc-42a6-b564-4fc470189f24-utilities\") pod \"c7ff8f3a-debc-42a6-b564-4fc470189f24\" (UID: \"c7ff8f3a-debc-42a6-b564-4fc470189f24\") " Dec 03 17:17:21 crc kubenswrapper[4998]: I1203 17:17:21.116529 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rj6b\" (UniqueName: \"kubernetes.io/projected/c7ff8f3a-debc-42a6-b564-4fc470189f24-kube-api-access-9rj6b\") pod \"c7ff8f3a-debc-42a6-b564-4fc470189f24\" (UID: \"c7ff8f3a-debc-42a6-b564-4fc470189f24\") " Dec 03 17:17:21 crc kubenswrapper[4998]: I1203 17:17:21.116550 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7ff8f3a-debc-42a6-b564-4fc470189f24-catalog-content\") pod \"c7ff8f3a-debc-42a6-b564-4fc470189f24\" (UID: \"c7ff8f3a-debc-42a6-b564-4fc470189f24\") " Dec 03 17:17:21 crc kubenswrapper[4998]: I1203 17:17:21.117167 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7ff8f3a-debc-42a6-b564-4fc470189f24-utilities" (OuterVolumeSpecName: "utilities") pod "c7ff8f3a-debc-42a6-b564-4fc470189f24" (UID: "c7ff8f3a-debc-42a6-b564-4fc470189f24"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:17:21 crc kubenswrapper[4998]: I1203 17:17:21.136957 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7ff8f3a-debc-42a6-b564-4fc470189f24-kube-api-access-9rj6b" (OuterVolumeSpecName: "kube-api-access-9rj6b") pod "c7ff8f3a-debc-42a6-b564-4fc470189f24" (UID: "c7ff8f3a-debc-42a6-b564-4fc470189f24"). InnerVolumeSpecName "kube-api-access-9rj6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:17:21 crc kubenswrapper[4998]: I1203 17:17:21.218469 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7ff8f3a-debc-42a6-b564-4fc470189f24-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:21 crc kubenswrapper[4998]: I1203 17:17:21.218502 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rj6b\" (UniqueName: \"kubernetes.io/projected/c7ff8f3a-debc-42a6-b564-4fc470189f24-kube-api-access-9rj6b\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:21 crc kubenswrapper[4998]: I1203 17:17:21.223230 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7ff8f3a-debc-42a6-b564-4fc470189f24-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c7ff8f3a-debc-42a6-b564-4fc470189f24" (UID: "c7ff8f3a-debc-42a6-b564-4fc470189f24"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:17:21 crc kubenswrapper[4998]: I1203 17:17:21.320125 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7ff8f3a-debc-42a6-b564-4fc470189f24-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:21 crc kubenswrapper[4998]: I1203 17:17:21.495701 4998 generic.go:334] "Generic (PLEG): container finished" podID="c7ff8f3a-debc-42a6-b564-4fc470189f24" containerID="6f9b1da1e59db40a0f75ec91965a14914e6d33d0752398dd61625a856ae20a3b" exitCode=0 Dec 03 17:17:21 crc kubenswrapper[4998]: I1203 17:17:21.495768 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4cmzb" event={"ID":"c7ff8f3a-debc-42a6-b564-4fc470189f24","Type":"ContainerDied","Data":"6f9b1da1e59db40a0f75ec91965a14914e6d33d0752398dd61625a856ae20a3b"} Dec 03 17:17:21 crc kubenswrapper[4998]: I1203 17:17:21.495803 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4cmzb" event={"ID":"c7ff8f3a-debc-42a6-b564-4fc470189f24","Type":"ContainerDied","Data":"3ee0364cffef6dcf72b5e9a22d7fa75eb0b809decdd9189e9183e1707141e8da"} Dec 03 17:17:21 crc kubenswrapper[4998]: I1203 17:17:21.495823 4998 scope.go:117] "RemoveContainer" containerID="6f9b1da1e59db40a0f75ec91965a14914e6d33d0752398dd61625a856ae20a3b" Dec 03 17:17:21 crc kubenswrapper[4998]: I1203 17:17:21.495830 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4cmzb" Dec 03 17:17:21 crc kubenswrapper[4998]: I1203 17:17:21.520056 4998 scope.go:117] "RemoveContainer" containerID="1082bebec6add1d63556b7af1d17f550cbd8ad09d6cf7f53bf5cda8f9fe25654" Dec 03 17:17:21 crc kubenswrapper[4998]: I1203 17:17:21.529735 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4cmzb"] Dec 03 17:17:21 crc kubenswrapper[4998]: I1203 17:17:21.541883 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4cmzb"] Dec 03 17:17:21 crc kubenswrapper[4998]: I1203 17:17:21.542498 4998 scope.go:117] "RemoveContainer" containerID="1da7a3c5628448b6e2a55db1b875092e2488adf9f80013ddc4347fd7e0847764" Dec 03 17:17:21 crc kubenswrapper[4998]: I1203 17:17:21.597639 4998 scope.go:117] "RemoveContainer" containerID="6f9b1da1e59db40a0f75ec91965a14914e6d33d0752398dd61625a856ae20a3b" Dec 03 17:17:21 crc kubenswrapper[4998]: E1203 17:17:21.598124 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f9b1da1e59db40a0f75ec91965a14914e6d33d0752398dd61625a856ae20a3b\": container with ID starting with 6f9b1da1e59db40a0f75ec91965a14914e6d33d0752398dd61625a856ae20a3b not found: ID does not exist" containerID="6f9b1da1e59db40a0f75ec91965a14914e6d33d0752398dd61625a856ae20a3b" Dec 03 17:17:21 crc kubenswrapper[4998]: I1203 17:17:21.598163 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f9b1da1e59db40a0f75ec91965a14914e6d33d0752398dd61625a856ae20a3b"} err="failed to get container status \"6f9b1da1e59db40a0f75ec91965a14914e6d33d0752398dd61625a856ae20a3b\": rpc error: code = NotFound desc = could not find container \"6f9b1da1e59db40a0f75ec91965a14914e6d33d0752398dd61625a856ae20a3b\": container with ID starting with 6f9b1da1e59db40a0f75ec91965a14914e6d33d0752398dd61625a856ae20a3b not found: ID does not exist" Dec 03 17:17:21 crc kubenswrapper[4998]: I1203 17:17:21.598184 4998 scope.go:117] "RemoveContainer" containerID="1082bebec6add1d63556b7af1d17f550cbd8ad09d6cf7f53bf5cda8f9fe25654" Dec 03 17:17:21 crc kubenswrapper[4998]: E1203 17:17:21.598412 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1082bebec6add1d63556b7af1d17f550cbd8ad09d6cf7f53bf5cda8f9fe25654\": container with ID starting with 1082bebec6add1d63556b7af1d17f550cbd8ad09d6cf7f53bf5cda8f9fe25654 not found: ID does not exist" containerID="1082bebec6add1d63556b7af1d17f550cbd8ad09d6cf7f53bf5cda8f9fe25654" Dec 03 17:17:21 crc kubenswrapper[4998]: I1203 17:17:21.598442 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1082bebec6add1d63556b7af1d17f550cbd8ad09d6cf7f53bf5cda8f9fe25654"} err="failed to get container status \"1082bebec6add1d63556b7af1d17f550cbd8ad09d6cf7f53bf5cda8f9fe25654\": rpc error: code = NotFound desc = could not find container \"1082bebec6add1d63556b7af1d17f550cbd8ad09d6cf7f53bf5cda8f9fe25654\": container with ID starting with 1082bebec6add1d63556b7af1d17f550cbd8ad09d6cf7f53bf5cda8f9fe25654 not found: ID does not exist" Dec 03 17:17:21 crc kubenswrapper[4998]: I1203 17:17:21.598473 4998 scope.go:117] "RemoveContainer" containerID="1da7a3c5628448b6e2a55db1b875092e2488adf9f80013ddc4347fd7e0847764" Dec 03 17:17:21 crc kubenswrapper[4998]: E1203 17:17:21.598803 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1da7a3c5628448b6e2a55db1b875092e2488adf9f80013ddc4347fd7e0847764\": container with ID starting with 1da7a3c5628448b6e2a55db1b875092e2488adf9f80013ddc4347fd7e0847764 not found: ID does not exist" containerID="1da7a3c5628448b6e2a55db1b875092e2488adf9f80013ddc4347fd7e0847764" Dec 03 17:17:21 crc kubenswrapper[4998]: I1203 17:17:21.598862 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1da7a3c5628448b6e2a55db1b875092e2488adf9f80013ddc4347fd7e0847764"} err="failed to get container status \"1da7a3c5628448b6e2a55db1b875092e2488adf9f80013ddc4347fd7e0847764\": rpc error: code = NotFound desc = could not find container \"1da7a3c5628448b6e2a55db1b875092e2488adf9f80013ddc4347fd7e0847764\": container with ID starting with 1da7a3c5628448b6e2a55db1b875092e2488adf9f80013ddc4347fd7e0847764 not found: ID does not exist" Dec 03 17:17:21 crc kubenswrapper[4998]: I1203 17:17:21.702635 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7ff8f3a-debc-42a6-b564-4fc470189f24" path="/var/lib/kubelet/pods/c7ff8f3a-debc-42a6-b564-4fc470189f24/volumes" Dec 03 17:17:29 crc kubenswrapper[4998]: I1203 17:17:29.678546 4998 scope.go:117] "RemoveContainer" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" Dec 03 17:17:29 crc kubenswrapper[4998]: E1203 17:17:29.679818 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:17:42 crc kubenswrapper[4998]: I1203 17:17:42.677728 4998 scope.go:117] "RemoveContainer" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" Dec 03 17:17:42 crc kubenswrapper[4998]: E1203 17:17:42.678619 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:17:47 crc kubenswrapper[4998]: I1203 17:17:47.902196 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-96jf4"] Dec 03 17:17:47 crc kubenswrapper[4998]: E1203 17:17:47.903203 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7ff8f3a-debc-42a6-b564-4fc470189f24" containerName="extract-utilities" Dec 03 17:17:47 crc kubenswrapper[4998]: I1203 17:17:47.903219 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7ff8f3a-debc-42a6-b564-4fc470189f24" containerName="extract-utilities" Dec 03 17:17:47 crc kubenswrapper[4998]: E1203 17:17:47.903231 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7ff8f3a-debc-42a6-b564-4fc470189f24" containerName="registry-server" Dec 03 17:17:47 crc kubenswrapper[4998]: I1203 17:17:47.903237 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7ff8f3a-debc-42a6-b564-4fc470189f24" containerName="registry-server" Dec 03 17:17:47 crc kubenswrapper[4998]: E1203 17:17:47.903272 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7ff8f3a-debc-42a6-b564-4fc470189f24" containerName="extract-content" Dec 03 17:17:47 crc kubenswrapper[4998]: I1203 17:17:47.903278 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7ff8f3a-debc-42a6-b564-4fc470189f24" containerName="extract-content" Dec 03 17:17:47 crc kubenswrapper[4998]: I1203 17:17:47.903485 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7ff8f3a-debc-42a6-b564-4fc470189f24" containerName="registry-server" Dec 03 17:17:47 crc kubenswrapper[4998]: I1203 17:17:47.905918 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-96jf4" Dec 03 17:17:47 crc kubenswrapper[4998]: I1203 17:17:47.929163 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-96jf4"] Dec 03 17:17:47 crc kubenswrapper[4998]: I1203 17:17:47.982056 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67-utilities\") pod \"redhat-marketplace-96jf4\" (UID: \"c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67\") " pod="openshift-marketplace/redhat-marketplace-96jf4" Dec 03 17:17:47 crc kubenswrapper[4998]: I1203 17:17:47.982142 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dslnx\" (UniqueName: \"kubernetes.io/projected/c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67-kube-api-access-dslnx\") pod \"redhat-marketplace-96jf4\" (UID: \"c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67\") " pod="openshift-marketplace/redhat-marketplace-96jf4" Dec 03 17:17:47 crc kubenswrapper[4998]: I1203 17:17:47.982394 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67-catalog-content\") pod \"redhat-marketplace-96jf4\" (UID: \"c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67\") " pod="openshift-marketplace/redhat-marketplace-96jf4" Dec 03 17:17:48 crc kubenswrapper[4998]: I1203 17:17:48.084787 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67-catalog-content\") pod \"redhat-marketplace-96jf4\" (UID: \"c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67\") " pod="openshift-marketplace/redhat-marketplace-96jf4" Dec 03 17:17:48 crc kubenswrapper[4998]: I1203 17:17:48.085144 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67-utilities\") pod \"redhat-marketplace-96jf4\" (UID: \"c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67\") " pod="openshift-marketplace/redhat-marketplace-96jf4" Dec 03 17:17:48 crc kubenswrapper[4998]: I1203 17:17:48.085197 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dslnx\" (UniqueName: \"kubernetes.io/projected/c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67-kube-api-access-dslnx\") pod \"redhat-marketplace-96jf4\" (UID: \"c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67\") " pod="openshift-marketplace/redhat-marketplace-96jf4" Dec 03 17:17:48 crc kubenswrapper[4998]: I1203 17:17:48.086283 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67-catalog-content\") pod \"redhat-marketplace-96jf4\" (UID: \"c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67\") " pod="openshift-marketplace/redhat-marketplace-96jf4" Dec 03 17:17:48 crc kubenswrapper[4998]: I1203 17:17:48.088006 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67-utilities\") pod \"redhat-marketplace-96jf4\" (UID: \"c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67\") " pod="openshift-marketplace/redhat-marketplace-96jf4" Dec 03 17:17:48 crc kubenswrapper[4998]: I1203 17:17:48.136959 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dslnx\" (UniqueName: \"kubernetes.io/projected/c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67-kube-api-access-dslnx\") pod \"redhat-marketplace-96jf4\" (UID: \"c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67\") " pod="openshift-marketplace/redhat-marketplace-96jf4" Dec 03 17:17:48 crc kubenswrapper[4998]: I1203 17:17:48.241293 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-96jf4" Dec 03 17:17:48 crc kubenswrapper[4998]: I1203 17:17:48.789683 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-96jf4"] Dec 03 17:17:49 crc kubenswrapper[4998]: I1203 17:17:49.781382 4998 generic.go:334] "Generic (PLEG): container finished" podID="c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67" containerID="ae878cd6cc683df6fee7a0d2dc33f563b4e8b69c16672f45b12f68d81d2d4cc1" exitCode=0 Dec 03 17:17:49 crc kubenswrapper[4998]: I1203 17:17:49.781962 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96jf4" event={"ID":"c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67","Type":"ContainerDied","Data":"ae878cd6cc683df6fee7a0d2dc33f563b4e8b69c16672f45b12f68d81d2d4cc1"} Dec 03 17:17:49 crc kubenswrapper[4998]: I1203 17:17:49.781992 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96jf4" event={"ID":"c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67","Type":"ContainerStarted","Data":"b9700feb067afb69917922a041f8aada94d169eba608fc4cbc0e08587beb1983"} Dec 03 17:17:50 crc kubenswrapper[4998]: I1203 17:17:50.792018 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96jf4" event={"ID":"c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67","Type":"ContainerStarted","Data":"ce7f13d9dc6738ed398267a0fee96d809ed854871dbf511a4f242ceea44c80e9"} Dec 03 17:17:51 crc kubenswrapper[4998]: I1203 17:17:51.813557 4998 generic.go:334] "Generic (PLEG): container finished" podID="c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67" containerID="ce7f13d9dc6738ed398267a0fee96d809ed854871dbf511a4f242ceea44c80e9" exitCode=0 Dec 03 17:17:51 crc kubenswrapper[4998]: I1203 17:17:51.813623 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96jf4" event={"ID":"c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67","Type":"ContainerDied","Data":"ce7f13d9dc6738ed398267a0fee96d809ed854871dbf511a4f242ceea44c80e9"} Dec 03 17:17:52 crc kubenswrapper[4998]: I1203 17:17:52.825285 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96jf4" event={"ID":"c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67","Type":"ContainerStarted","Data":"230d93b46504c22b5d4d99aba482d383953a936dae99fd217264d3bbdd33b6c1"} Dec 03 17:17:52 crc kubenswrapper[4998]: I1203 17:17:52.848055 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-96jf4" podStartSLOduration=3.408914241 podStartE2EDuration="5.848032837s" podCreationTimestamp="2025-12-03 17:17:47 +0000 UTC" firstStartedPulling="2025-12-03 17:17:49.784335972 +0000 UTC m=+4448.396036195" lastFinishedPulling="2025-12-03 17:17:52.223454568 +0000 UTC m=+4450.835154791" observedRunningTime="2025-12-03 17:17:52.846679593 +0000 UTC m=+4451.458379846" watchObservedRunningTime="2025-12-03 17:17:52.848032837 +0000 UTC m=+4451.459733060" Dec 03 17:17:57 crc kubenswrapper[4998]: I1203 17:17:57.680415 4998 scope.go:117] "RemoveContainer" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" Dec 03 17:17:57 crc kubenswrapper[4998]: E1203 17:17:57.681030 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:17:58 crc kubenswrapper[4998]: I1203 17:17:58.243469 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-96jf4" Dec 03 17:17:58 crc kubenswrapper[4998]: I1203 17:17:58.243541 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-96jf4" Dec 03 17:17:58 crc kubenswrapper[4998]: I1203 17:17:58.300859 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-96jf4" Dec 03 17:17:58 crc kubenswrapper[4998]: I1203 17:17:58.935981 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-96jf4" Dec 03 17:17:58 crc kubenswrapper[4998]: I1203 17:17:58.982681 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-96jf4"] Dec 03 17:18:00 crc kubenswrapper[4998]: I1203 17:18:00.895929 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-96jf4" podUID="c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67" containerName="registry-server" containerID="cri-o://230d93b46504c22b5d4d99aba482d383953a936dae99fd217264d3bbdd33b6c1" gracePeriod=2 Dec 03 17:18:01 crc kubenswrapper[4998]: I1203 17:18:01.907499 4998 generic.go:334] "Generic (PLEG): container finished" podID="c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67" containerID="230d93b46504c22b5d4d99aba482d383953a936dae99fd217264d3bbdd33b6c1" exitCode=0 Dec 03 17:18:01 crc kubenswrapper[4998]: I1203 17:18:01.907587 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96jf4" event={"ID":"c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67","Type":"ContainerDied","Data":"230d93b46504c22b5d4d99aba482d383953a936dae99fd217264d3bbdd33b6c1"} Dec 03 17:18:03 crc kubenswrapper[4998]: I1203 17:18:03.112036 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-96jf4" Dec 03 17:18:03 crc kubenswrapper[4998]: I1203 17:18:03.308705 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67-utilities\") pod \"c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67\" (UID: \"c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67\") " Dec 03 17:18:03 crc kubenswrapper[4998]: I1203 17:18:03.309217 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67-catalog-content\") pod \"c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67\" (UID: \"c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67\") " Dec 03 17:18:03 crc kubenswrapper[4998]: I1203 17:18:03.309552 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dslnx\" (UniqueName: \"kubernetes.io/projected/c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67-kube-api-access-dslnx\") pod \"c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67\" (UID: \"c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67\") " Dec 03 17:18:03 crc kubenswrapper[4998]: I1203 17:18:03.309558 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67-utilities" (OuterVolumeSpecName: "utilities") pod "c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67" (UID: "c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:18:03 crc kubenswrapper[4998]: I1203 17:18:03.310192 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:03 crc kubenswrapper[4998]: I1203 17:18:03.319159 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67-kube-api-access-dslnx" (OuterVolumeSpecName: "kube-api-access-dslnx") pod "c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67" (UID: "c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67"). InnerVolumeSpecName "kube-api-access-dslnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:18:03 crc kubenswrapper[4998]: I1203 17:18:03.328716 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67" (UID: "c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:18:03 crc kubenswrapper[4998]: I1203 17:18:03.412183 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dslnx\" (UniqueName: \"kubernetes.io/projected/c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67-kube-api-access-dslnx\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:03 crc kubenswrapper[4998]: I1203 17:18:03.412236 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:03 crc kubenswrapper[4998]: I1203 17:18:03.959534 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96jf4" event={"ID":"c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67","Type":"ContainerDied","Data":"b9700feb067afb69917922a041f8aada94d169eba608fc4cbc0e08587beb1983"} Dec 03 17:18:03 crc kubenswrapper[4998]: I1203 17:18:03.959608 4998 scope.go:117] "RemoveContainer" containerID="230d93b46504c22b5d4d99aba482d383953a936dae99fd217264d3bbdd33b6c1" Dec 03 17:18:03 crc kubenswrapper[4998]: I1203 17:18:03.959607 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-96jf4" Dec 03 17:18:04 crc kubenswrapper[4998]: I1203 17:18:04.022313 4998 scope.go:117] "RemoveContainer" containerID="ce7f13d9dc6738ed398267a0fee96d809ed854871dbf511a4f242ceea44c80e9" Dec 03 17:18:04 crc kubenswrapper[4998]: I1203 17:18:04.076919 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-96jf4"] Dec 03 17:18:04 crc kubenswrapper[4998]: I1203 17:18:04.098018 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-96jf4"] Dec 03 17:18:04 crc kubenswrapper[4998]: I1203 17:18:04.121295 4998 scope.go:117] "RemoveContainer" containerID="ae878cd6cc683df6fee7a0d2dc33f563b4e8b69c16672f45b12f68d81d2d4cc1" Dec 03 17:18:05 crc kubenswrapper[4998]: I1203 17:18:05.690312 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67" path="/var/lib/kubelet/pods/c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67/volumes" Dec 03 17:18:12 crc kubenswrapper[4998]: I1203 17:18:12.678386 4998 scope.go:117] "RemoveContainer" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" Dec 03 17:18:12 crc kubenswrapper[4998]: E1203 17:18:12.679561 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:18:27 crc kubenswrapper[4998]: I1203 17:18:27.678892 4998 scope.go:117] "RemoveContainer" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" Dec 03 17:18:27 crc kubenswrapper[4998]: E1203 17:18:27.679956 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:18:38 crc kubenswrapper[4998]: I1203 17:18:38.677996 4998 scope.go:117] "RemoveContainer" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" Dec 03 17:18:38 crc kubenswrapper[4998]: E1203 17:18:38.679010 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:18:50 crc kubenswrapper[4998]: I1203 17:18:50.678226 4998 scope.go:117] "RemoveContainer" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" Dec 03 17:18:50 crc kubenswrapper[4998]: E1203 17:18:50.678996 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:19:02 crc kubenswrapper[4998]: I1203 17:19:02.678798 4998 scope.go:117] "RemoveContainer" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" Dec 03 17:19:02 crc kubenswrapper[4998]: E1203 17:19:02.679638 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:19:14 crc kubenswrapper[4998]: I1203 17:19:14.678050 4998 scope.go:117] "RemoveContainer" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" Dec 03 17:19:14 crc kubenswrapper[4998]: E1203 17:19:14.679168 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:19:27 crc kubenswrapper[4998]: I1203 17:19:27.681066 4998 scope.go:117] "RemoveContainer" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" Dec 03 17:19:28 crc kubenswrapper[4998]: I1203 17:19:28.892976 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"e41ffa6ccaf21547e37ed2928d4106bde8dad9ea46462bf6f61fce217f12ad7d"} Dec 03 17:20:02 crc kubenswrapper[4998]: I1203 17:20:02.094511 4998 scope.go:117] "RemoveContainer" containerID="5c382418cd117cfbb9dffafdc02c664699f910089593923cadbcec38eca24195" Dec 03 17:20:02 crc kubenswrapper[4998]: I1203 17:20:02.129344 4998 scope.go:117] "RemoveContainer" containerID="d9852c7ac707d915cd27694fca3c394d14eb1eea1d29b6e98db1b9dee49b3dd8" Dec 03 17:20:02 crc kubenswrapper[4998]: I1203 17:20:02.186084 4998 scope.go:117] "RemoveContainer" containerID="d7c60711e6baa20508897887238f9adcb6513ff75c7abac8cca54b5acdd202f7" Dec 03 17:21:27 crc kubenswrapper[4998]: I1203 17:21:27.111332 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:21:27 crc kubenswrapper[4998]: I1203 17:21:27.112211 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:21:57 crc kubenswrapper[4998]: I1203 17:21:57.110769 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:21:57 crc kubenswrapper[4998]: I1203 17:21:57.111322 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:22:27 crc kubenswrapper[4998]: I1203 17:22:27.111361 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:22:27 crc kubenswrapper[4998]: I1203 17:22:27.112027 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:22:27 crc kubenswrapper[4998]: I1203 17:22:27.112124 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 17:22:27 crc kubenswrapper[4998]: I1203 17:22:27.113675 4998 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e41ffa6ccaf21547e37ed2928d4106bde8dad9ea46462bf6f61fce217f12ad7d"} pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:22:27 crc kubenswrapper[4998]: I1203 17:22:27.113847 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" containerID="cri-o://e41ffa6ccaf21547e37ed2928d4106bde8dad9ea46462bf6f61fce217f12ad7d" gracePeriod=600 Dec 03 17:22:27 crc kubenswrapper[4998]: I1203 17:22:27.388190 4998 generic.go:334] "Generic (PLEG): container finished" podID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerID="e41ffa6ccaf21547e37ed2928d4106bde8dad9ea46462bf6f61fce217f12ad7d" exitCode=0 Dec 03 17:22:27 crc kubenswrapper[4998]: I1203 17:22:27.388247 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerDied","Data":"e41ffa6ccaf21547e37ed2928d4106bde8dad9ea46462bf6f61fce217f12ad7d"} Dec 03 17:22:27 crc kubenswrapper[4998]: I1203 17:22:27.388292 4998 scope.go:117] "RemoveContainer" containerID="3306c9359d37b408be08c568e2caade0af42ad04872a4cd90742f591b9c53a85" Dec 03 17:22:27 crc kubenswrapper[4998]: E1203 17:22:27.445499 4998 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27aeb142_d4e3_4827_ac1f_c34de6822b14.slice/crio-e41ffa6ccaf21547e37ed2928d4106bde8dad9ea46462bf6f61fce217f12ad7d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27aeb142_d4e3_4827_ac1f_c34de6822b14.slice/crio-conmon-e41ffa6ccaf21547e37ed2928d4106bde8dad9ea46462bf6f61fce217f12ad7d.scope\": RecentStats: unable to find data in memory cache]" Dec 03 17:22:28 crc kubenswrapper[4998]: I1203 17:22:28.404663 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00"} Dec 03 17:24:27 crc kubenswrapper[4998]: I1203 17:24:27.111398 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:24:27 crc kubenswrapper[4998]: I1203 17:24:27.112101 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:24:30 crc kubenswrapper[4998]: I1203 17:24:30.211642 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vf7bk"] Dec 03 17:24:30 crc kubenswrapper[4998]: E1203 17:24:30.212883 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67" containerName="extract-content" Dec 03 17:24:30 crc kubenswrapper[4998]: I1203 17:24:30.212906 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67" containerName="extract-content" Dec 03 17:24:30 crc kubenswrapper[4998]: E1203 17:24:30.212928 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67" containerName="extract-utilities" Dec 03 17:24:30 crc kubenswrapper[4998]: I1203 17:24:30.212939 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67" containerName="extract-utilities" Dec 03 17:24:30 crc kubenswrapper[4998]: E1203 17:24:30.212978 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67" containerName="registry-server" Dec 03 17:24:30 crc kubenswrapper[4998]: I1203 17:24:30.212990 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67" containerName="registry-server" Dec 03 17:24:30 crc kubenswrapper[4998]: I1203 17:24:30.213324 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="c15b9d9a-0ef6-41c5-bf67-40fbd9bf8f67" containerName="registry-server" Dec 03 17:24:30 crc kubenswrapper[4998]: I1203 17:24:30.215601 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vf7bk" Dec 03 17:24:30 crc kubenswrapper[4998]: I1203 17:24:30.231963 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vf7bk"] Dec 03 17:24:30 crc kubenswrapper[4998]: I1203 17:24:30.345735 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/391a6994-a68d-49df-aa69-22bc86591072-utilities\") pod \"redhat-operators-vf7bk\" (UID: \"391a6994-a68d-49df-aa69-22bc86591072\") " pod="openshift-marketplace/redhat-operators-vf7bk" Dec 03 17:24:30 crc kubenswrapper[4998]: I1203 17:24:30.345868 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh8cp\" (UniqueName: \"kubernetes.io/projected/391a6994-a68d-49df-aa69-22bc86591072-kube-api-access-jh8cp\") pod \"redhat-operators-vf7bk\" (UID: \"391a6994-a68d-49df-aa69-22bc86591072\") " pod="openshift-marketplace/redhat-operators-vf7bk" Dec 03 17:24:30 crc kubenswrapper[4998]: I1203 17:24:30.345979 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/391a6994-a68d-49df-aa69-22bc86591072-catalog-content\") pod \"redhat-operators-vf7bk\" (UID: \"391a6994-a68d-49df-aa69-22bc86591072\") " pod="openshift-marketplace/redhat-operators-vf7bk" Dec 03 17:24:30 crc kubenswrapper[4998]: I1203 17:24:30.448404 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/391a6994-a68d-49df-aa69-22bc86591072-catalog-content\") pod \"redhat-operators-vf7bk\" (UID: \"391a6994-a68d-49df-aa69-22bc86591072\") " pod="openshift-marketplace/redhat-operators-vf7bk" Dec 03 17:24:30 crc kubenswrapper[4998]: I1203 17:24:30.448579 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/391a6994-a68d-49df-aa69-22bc86591072-utilities\") pod \"redhat-operators-vf7bk\" (UID: \"391a6994-a68d-49df-aa69-22bc86591072\") " pod="openshift-marketplace/redhat-operators-vf7bk" Dec 03 17:24:30 crc kubenswrapper[4998]: I1203 17:24:30.448715 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh8cp\" (UniqueName: \"kubernetes.io/projected/391a6994-a68d-49df-aa69-22bc86591072-kube-api-access-jh8cp\") pod \"redhat-operators-vf7bk\" (UID: \"391a6994-a68d-49df-aa69-22bc86591072\") " pod="openshift-marketplace/redhat-operators-vf7bk" Dec 03 17:24:30 crc kubenswrapper[4998]: I1203 17:24:30.449105 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/391a6994-a68d-49df-aa69-22bc86591072-catalog-content\") pod \"redhat-operators-vf7bk\" (UID: \"391a6994-a68d-49df-aa69-22bc86591072\") " pod="openshift-marketplace/redhat-operators-vf7bk" Dec 03 17:24:30 crc kubenswrapper[4998]: I1203 17:24:30.449238 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/391a6994-a68d-49df-aa69-22bc86591072-utilities\") pod \"redhat-operators-vf7bk\" (UID: \"391a6994-a68d-49df-aa69-22bc86591072\") " pod="openshift-marketplace/redhat-operators-vf7bk" Dec 03 17:24:30 crc kubenswrapper[4998]: I1203 17:24:30.482107 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh8cp\" (UniqueName: \"kubernetes.io/projected/391a6994-a68d-49df-aa69-22bc86591072-kube-api-access-jh8cp\") pod \"redhat-operators-vf7bk\" (UID: \"391a6994-a68d-49df-aa69-22bc86591072\") " pod="openshift-marketplace/redhat-operators-vf7bk" Dec 03 17:24:30 crc kubenswrapper[4998]: I1203 17:24:30.536821 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vf7bk" Dec 03 17:24:31 crc kubenswrapper[4998]: I1203 17:24:31.040731 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vf7bk"] Dec 03 17:24:31 crc kubenswrapper[4998]: I1203 17:24:31.886795 4998 generic.go:334] "Generic (PLEG): container finished" podID="391a6994-a68d-49df-aa69-22bc86591072" containerID="a0aa970f180fd5806ea01288c748f7ee62468e676801fe71a23757bf364c2831" exitCode=0 Dec 03 17:24:31 crc kubenswrapper[4998]: I1203 17:24:31.887364 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vf7bk" event={"ID":"391a6994-a68d-49df-aa69-22bc86591072","Type":"ContainerDied","Data":"a0aa970f180fd5806ea01288c748f7ee62468e676801fe71a23757bf364c2831"} Dec 03 17:24:31 crc kubenswrapper[4998]: I1203 17:24:31.887395 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vf7bk" event={"ID":"391a6994-a68d-49df-aa69-22bc86591072","Type":"ContainerStarted","Data":"e38fb9c88f64a2f87cc38eb0d96bdc412ce66f244bb6699875c74208a2cf4ef1"} Dec 03 17:24:31 crc kubenswrapper[4998]: I1203 17:24:31.889733 4998 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 17:24:33 crc kubenswrapper[4998]: I1203 17:24:33.910922 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vf7bk" event={"ID":"391a6994-a68d-49df-aa69-22bc86591072","Type":"ContainerStarted","Data":"4e16e090c8027a2f87de895edc3762c0512bcab99ca42a588da436a1588d0ee3"} Dec 03 17:24:36 crc kubenswrapper[4998]: I1203 17:24:36.966612 4998 generic.go:334] "Generic (PLEG): container finished" podID="391a6994-a68d-49df-aa69-22bc86591072" containerID="4e16e090c8027a2f87de895edc3762c0512bcab99ca42a588da436a1588d0ee3" exitCode=0 Dec 03 17:24:36 crc kubenswrapper[4998]: I1203 17:24:36.966694 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vf7bk" event={"ID":"391a6994-a68d-49df-aa69-22bc86591072","Type":"ContainerDied","Data":"4e16e090c8027a2f87de895edc3762c0512bcab99ca42a588da436a1588d0ee3"} Dec 03 17:24:37 crc kubenswrapper[4998]: I1203 17:24:37.985132 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vf7bk" event={"ID":"391a6994-a68d-49df-aa69-22bc86591072","Type":"ContainerStarted","Data":"b44608923bd24d879b1229a0789b2274dcfdfec8ca09bbca064bf70453f7d22f"} Dec 03 17:24:38 crc kubenswrapper[4998]: I1203 17:24:38.025434 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vf7bk" podStartSLOduration=2.547921709 podStartE2EDuration="8.025408545s" podCreationTimestamp="2025-12-03 17:24:30 +0000 UTC" firstStartedPulling="2025-12-03 17:24:31.889256031 +0000 UTC m=+4850.500956294" lastFinishedPulling="2025-12-03 17:24:37.366742907 +0000 UTC m=+4855.978443130" observedRunningTime="2025-12-03 17:24:38.013050436 +0000 UTC m=+4856.624750679" watchObservedRunningTime="2025-12-03 17:24:38.025408545 +0000 UTC m=+4856.637108808" Dec 03 17:24:40 crc kubenswrapper[4998]: I1203 17:24:40.542495 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vf7bk" Dec 03 17:24:40 crc kubenswrapper[4998]: I1203 17:24:40.545384 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vf7bk" Dec 03 17:24:41 crc kubenswrapper[4998]: I1203 17:24:41.596541 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vf7bk" podUID="391a6994-a68d-49df-aa69-22bc86591072" containerName="registry-server" probeResult="failure" output=< Dec 03 17:24:41 crc kubenswrapper[4998]: timeout: failed to connect service ":50051" within 1s Dec 03 17:24:41 crc kubenswrapper[4998]: > Dec 03 17:24:50 crc kubenswrapper[4998]: I1203 17:24:50.609066 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vf7bk" Dec 03 17:24:50 crc kubenswrapper[4998]: I1203 17:24:50.668050 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vf7bk" Dec 03 17:24:50 crc kubenswrapper[4998]: I1203 17:24:50.857640 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vf7bk"] Dec 03 17:24:52 crc kubenswrapper[4998]: I1203 17:24:52.128419 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vf7bk" podUID="391a6994-a68d-49df-aa69-22bc86591072" containerName="registry-server" containerID="cri-o://b44608923bd24d879b1229a0789b2274dcfdfec8ca09bbca064bf70453f7d22f" gracePeriod=2 Dec 03 17:24:52 crc kubenswrapper[4998]: I1203 17:24:52.607418 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vf7bk" Dec 03 17:24:52 crc kubenswrapper[4998]: I1203 17:24:52.795273 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/391a6994-a68d-49df-aa69-22bc86591072-utilities\") pod \"391a6994-a68d-49df-aa69-22bc86591072\" (UID: \"391a6994-a68d-49df-aa69-22bc86591072\") " Dec 03 17:24:52 crc kubenswrapper[4998]: I1203 17:24:52.795678 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jh8cp\" (UniqueName: \"kubernetes.io/projected/391a6994-a68d-49df-aa69-22bc86591072-kube-api-access-jh8cp\") pod \"391a6994-a68d-49df-aa69-22bc86591072\" (UID: \"391a6994-a68d-49df-aa69-22bc86591072\") " Dec 03 17:24:52 crc kubenswrapper[4998]: I1203 17:24:52.795769 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/391a6994-a68d-49df-aa69-22bc86591072-catalog-content\") pod \"391a6994-a68d-49df-aa69-22bc86591072\" (UID: \"391a6994-a68d-49df-aa69-22bc86591072\") " Dec 03 17:24:52 crc kubenswrapper[4998]: I1203 17:24:52.796623 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/391a6994-a68d-49df-aa69-22bc86591072-utilities" (OuterVolumeSpecName: "utilities") pod "391a6994-a68d-49df-aa69-22bc86591072" (UID: "391a6994-a68d-49df-aa69-22bc86591072"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:24:52 crc kubenswrapper[4998]: I1203 17:24:52.796900 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/391a6994-a68d-49df-aa69-22bc86591072-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:24:52 crc kubenswrapper[4998]: I1203 17:24:52.917480 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/391a6994-a68d-49df-aa69-22bc86591072-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "391a6994-a68d-49df-aa69-22bc86591072" (UID: "391a6994-a68d-49df-aa69-22bc86591072"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:24:53 crc kubenswrapper[4998]: I1203 17:24:53.660165 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/391a6994-a68d-49df-aa69-22bc86591072-kube-api-access-jh8cp" (OuterVolumeSpecName: "kube-api-access-jh8cp") pod "391a6994-a68d-49df-aa69-22bc86591072" (UID: "391a6994-a68d-49df-aa69-22bc86591072"). InnerVolumeSpecName "kube-api-access-jh8cp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:24:53 crc kubenswrapper[4998]: I1203 17:24:53.676580 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jh8cp\" (UniqueName: \"kubernetes.io/projected/391a6994-a68d-49df-aa69-22bc86591072-kube-api-access-jh8cp\") on node \"crc\" DevicePath \"\"" Dec 03 17:24:53 crc kubenswrapper[4998]: I1203 17:24:53.676625 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/391a6994-a68d-49df-aa69-22bc86591072-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:24:53 crc kubenswrapper[4998]: I1203 17:24:53.690064 4998 generic.go:334] "Generic (PLEG): container finished" podID="391a6994-a68d-49df-aa69-22bc86591072" containerID="b44608923bd24d879b1229a0789b2274dcfdfec8ca09bbca064bf70453f7d22f" exitCode=0 Dec 03 17:24:53 crc kubenswrapper[4998]: I1203 17:24:53.690164 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vf7bk" Dec 03 17:24:53 crc kubenswrapper[4998]: I1203 17:24:53.694918 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vf7bk" event={"ID":"391a6994-a68d-49df-aa69-22bc86591072","Type":"ContainerDied","Data":"b44608923bd24d879b1229a0789b2274dcfdfec8ca09bbca064bf70453f7d22f"} Dec 03 17:24:53 crc kubenswrapper[4998]: I1203 17:24:53.694967 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vf7bk" event={"ID":"391a6994-a68d-49df-aa69-22bc86591072","Type":"ContainerDied","Data":"e38fb9c88f64a2f87cc38eb0d96bdc412ce66f244bb6699875c74208a2cf4ef1"} Dec 03 17:24:53 crc kubenswrapper[4998]: I1203 17:24:53.695003 4998 scope.go:117] "RemoveContainer" containerID="b44608923bd24d879b1229a0789b2274dcfdfec8ca09bbca064bf70453f7d22f" Dec 03 17:24:53 crc kubenswrapper[4998]: I1203 17:24:53.719657 4998 scope.go:117] "RemoveContainer" containerID="4e16e090c8027a2f87de895edc3762c0512bcab99ca42a588da436a1588d0ee3" Dec 03 17:24:53 crc kubenswrapper[4998]: I1203 17:24:53.757208 4998 scope.go:117] "RemoveContainer" containerID="a0aa970f180fd5806ea01288c748f7ee62468e676801fe71a23757bf364c2831" Dec 03 17:24:53 crc kubenswrapper[4998]: I1203 17:24:53.813908 4998 scope.go:117] "RemoveContainer" containerID="b44608923bd24d879b1229a0789b2274dcfdfec8ca09bbca064bf70453f7d22f" Dec 03 17:24:53 crc kubenswrapper[4998]: E1203 17:24:53.814367 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b44608923bd24d879b1229a0789b2274dcfdfec8ca09bbca064bf70453f7d22f\": container with ID starting with b44608923bd24d879b1229a0789b2274dcfdfec8ca09bbca064bf70453f7d22f not found: ID does not exist" containerID="b44608923bd24d879b1229a0789b2274dcfdfec8ca09bbca064bf70453f7d22f" Dec 03 17:24:53 crc kubenswrapper[4998]: I1203 17:24:53.814405 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b44608923bd24d879b1229a0789b2274dcfdfec8ca09bbca064bf70453f7d22f"} err="failed to get container status \"b44608923bd24d879b1229a0789b2274dcfdfec8ca09bbca064bf70453f7d22f\": rpc error: code = NotFound desc = could not find container \"b44608923bd24d879b1229a0789b2274dcfdfec8ca09bbca064bf70453f7d22f\": container with ID starting with b44608923bd24d879b1229a0789b2274dcfdfec8ca09bbca064bf70453f7d22f not found: ID does not exist" Dec 03 17:24:53 crc kubenswrapper[4998]: I1203 17:24:53.814432 4998 scope.go:117] "RemoveContainer" containerID="4e16e090c8027a2f87de895edc3762c0512bcab99ca42a588da436a1588d0ee3" Dec 03 17:24:53 crc kubenswrapper[4998]: E1203 17:24:53.814711 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e16e090c8027a2f87de895edc3762c0512bcab99ca42a588da436a1588d0ee3\": container with ID starting with 4e16e090c8027a2f87de895edc3762c0512bcab99ca42a588da436a1588d0ee3 not found: ID does not exist" containerID="4e16e090c8027a2f87de895edc3762c0512bcab99ca42a588da436a1588d0ee3" Dec 03 17:24:53 crc kubenswrapper[4998]: I1203 17:24:53.814738 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e16e090c8027a2f87de895edc3762c0512bcab99ca42a588da436a1588d0ee3"} err="failed to get container status \"4e16e090c8027a2f87de895edc3762c0512bcab99ca42a588da436a1588d0ee3\": rpc error: code = NotFound desc = could not find container \"4e16e090c8027a2f87de895edc3762c0512bcab99ca42a588da436a1588d0ee3\": container with ID starting with 4e16e090c8027a2f87de895edc3762c0512bcab99ca42a588da436a1588d0ee3 not found: ID does not exist" Dec 03 17:24:53 crc kubenswrapper[4998]: I1203 17:24:53.814772 4998 scope.go:117] "RemoveContainer" containerID="a0aa970f180fd5806ea01288c748f7ee62468e676801fe71a23757bf364c2831" Dec 03 17:24:53 crc kubenswrapper[4998]: E1203 17:24:53.815015 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0aa970f180fd5806ea01288c748f7ee62468e676801fe71a23757bf364c2831\": container with ID starting with a0aa970f180fd5806ea01288c748f7ee62468e676801fe71a23757bf364c2831 not found: ID does not exist" containerID="a0aa970f180fd5806ea01288c748f7ee62468e676801fe71a23757bf364c2831" Dec 03 17:24:53 crc kubenswrapper[4998]: I1203 17:24:53.815043 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0aa970f180fd5806ea01288c748f7ee62468e676801fe71a23757bf364c2831"} err="failed to get container status \"a0aa970f180fd5806ea01288c748f7ee62468e676801fe71a23757bf364c2831\": rpc error: code = NotFound desc = could not find container \"a0aa970f180fd5806ea01288c748f7ee62468e676801fe71a23757bf364c2831\": container with ID starting with a0aa970f180fd5806ea01288c748f7ee62468e676801fe71a23757bf364c2831 not found: ID does not exist" Dec 03 17:24:57 crc kubenswrapper[4998]: I1203 17:24:57.111507 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:24:57 crc kubenswrapper[4998]: I1203 17:24:57.112349 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:25:23 crc kubenswrapper[4998]: I1203 17:25:23.706043 4998 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","burstable","pod391a6994-a68d-49df-aa69-22bc86591072"] err="unable to destroy cgroup paths for cgroup [kubepods burstable pod391a6994-a68d-49df-aa69-22bc86591072] : Timed out while waiting for systemd to remove kubepods-burstable-pod391a6994_a68d_49df_aa69_22bc86591072.slice" Dec 03 17:25:23 crc kubenswrapper[4998]: E1203 17:25:23.706955 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods burstable pod391a6994-a68d-49df-aa69-22bc86591072] : unable to destroy cgroup paths for cgroup [kubepods burstable pod391a6994-a68d-49df-aa69-22bc86591072] : Timed out while waiting for systemd to remove kubepods-burstable-pod391a6994_a68d_49df_aa69_22bc86591072.slice" pod="openshift-marketplace/redhat-operators-vf7bk" podUID="391a6994-a68d-49df-aa69-22bc86591072" Dec 03 17:25:24 crc kubenswrapper[4998]: I1203 17:25:24.082100 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vf7bk" Dec 03 17:25:24 crc kubenswrapper[4998]: I1203 17:25:24.152470 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vf7bk"] Dec 03 17:25:24 crc kubenswrapper[4998]: I1203 17:25:24.172897 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vf7bk"] Dec 03 17:25:25 crc kubenswrapper[4998]: I1203 17:25:25.691065 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="391a6994-a68d-49df-aa69-22bc86591072" path="/var/lib/kubelet/pods/391a6994-a68d-49df-aa69-22bc86591072/volumes" Dec 03 17:25:27 crc kubenswrapper[4998]: I1203 17:25:27.110868 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:25:27 crc kubenswrapper[4998]: I1203 17:25:27.111618 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:25:27 crc kubenswrapper[4998]: I1203 17:25:27.111681 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 17:25:27 crc kubenswrapper[4998]: I1203 17:25:27.112578 4998 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00"} pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:25:27 crc kubenswrapper[4998]: I1203 17:25:27.112660 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" containerID="cri-o://b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" gracePeriod=600 Dec 03 17:25:27 crc kubenswrapper[4998]: E1203 17:25:27.239629 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:25:28 crc kubenswrapper[4998]: I1203 17:25:28.131208 4998 generic.go:334] "Generic (PLEG): container finished" podID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" exitCode=0 Dec 03 17:25:28 crc kubenswrapper[4998]: I1203 17:25:28.131336 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerDied","Data":"b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00"} Dec 03 17:25:28 crc kubenswrapper[4998]: I1203 17:25:28.131722 4998 scope.go:117] "RemoveContainer" containerID="e41ffa6ccaf21547e37ed2928d4106bde8dad9ea46462bf6f61fce217f12ad7d" Dec 03 17:25:28 crc kubenswrapper[4998]: I1203 17:25:28.132792 4998 scope.go:117] "RemoveContainer" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" Dec 03 17:25:28 crc kubenswrapper[4998]: E1203 17:25:28.133498 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:25:42 crc kubenswrapper[4998]: I1203 17:25:42.678971 4998 scope.go:117] "RemoveContainer" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" Dec 03 17:25:42 crc kubenswrapper[4998]: E1203 17:25:42.680193 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:25:57 crc kubenswrapper[4998]: I1203 17:25:57.678490 4998 scope.go:117] "RemoveContainer" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" Dec 03 17:25:57 crc kubenswrapper[4998]: E1203 17:25:57.679501 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:26:11 crc kubenswrapper[4998]: I1203 17:26:11.689215 4998 scope.go:117] "RemoveContainer" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" Dec 03 17:26:11 crc kubenswrapper[4998]: E1203 17:26:11.690046 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:26:23 crc kubenswrapper[4998]: I1203 17:26:23.678428 4998 scope.go:117] "RemoveContainer" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" Dec 03 17:26:23 crc kubenswrapper[4998]: E1203 17:26:23.679736 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:26:36 crc kubenswrapper[4998]: I1203 17:26:36.679480 4998 scope.go:117] "RemoveContainer" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" Dec 03 17:26:36 crc kubenswrapper[4998]: E1203 17:26:36.680504 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:26:47 crc kubenswrapper[4998]: I1203 17:26:47.678075 4998 scope.go:117] "RemoveContainer" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" Dec 03 17:26:47 crc kubenswrapper[4998]: E1203 17:26:47.679100 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:27:00 crc kubenswrapper[4998]: I1203 17:27:00.678416 4998 scope.go:117] "RemoveContainer" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" Dec 03 17:27:00 crc kubenswrapper[4998]: E1203 17:27:00.679568 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:27:11 crc kubenswrapper[4998]: I1203 17:27:11.693214 4998 scope.go:117] "RemoveContainer" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" Dec 03 17:27:11 crc kubenswrapper[4998]: E1203 17:27:11.695849 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:27:26 crc kubenswrapper[4998]: I1203 17:27:26.677700 4998 scope.go:117] "RemoveContainer" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" Dec 03 17:27:26 crc kubenswrapper[4998]: E1203 17:27:26.678652 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:27:41 crc kubenswrapper[4998]: I1203 17:27:41.688261 4998 scope.go:117] "RemoveContainer" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" Dec 03 17:27:41 crc kubenswrapper[4998]: E1203 17:27:41.689743 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:27:55 crc kubenswrapper[4998]: I1203 17:27:55.682219 4998 scope.go:117] "RemoveContainer" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" Dec 03 17:27:55 crc kubenswrapper[4998]: E1203 17:27:55.682997 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:28:09 crc kubenswrapper[4998]: I1203 17:28:09.573256 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mpzr9"] Dec 03 17:28:09 crc kubenswrapper[4998]: E1203 17:28:09.574174 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="391a6994-a68d-49df-aa69-22bc86591072" containerName="extract-utilities" Dec 03 17:28:09 crc kubenswrapper[4998]: I1203 17:28:09.574188 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="391a6994-a68d-49df-aa69-22bc86591072" containerName="extract-utilities" Dec 03 17:28:09 crc kubenswrapper[4998]: E1203 17:28:09.574199 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="391a6994-a68d-49df-aa69-22bc86591072" containerName="registry-server" Dec 03 17:28:09 crc kubenswrapper[4998]: I1203 17:28:09.574205 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="391a6994-a68d-49df-aa69-22bc86591072" containerName="registry-server" Dec 03 17:28:09 crc kubenswrapper[4998]: E1203 17:28:09.574230 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="391a6994-a68d-49df-aa69-22bc86591072" containerName="extract-content" Dec 03 17:28:09 crc kubenswrapper[4998]: I1203 17:28:09.574237 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="391a6994-a68d-49df-aa69-22bc86591072" containerName="extract-content" Dec 03 17:28:09 crc kubenswrapper[4998]: I1203 17:28:09.574440 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="391a6994-a68d-49df-aa69-22bc86591072" containerName="registry-server" Dec 03 17:28:09 crc kubenswrapper[4998]: I1203 17:28:09.578474 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mpzr9" Dec 03 17:28:09 crc kubenswrapper[4998]: I1203 17:28:09.594978 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xblmg\" (UniqueName: \"kubernetes.io/projected/711fb3fa-94e5-499d-b1f5-ca2fe7326310-kube-api-access-xblmg\") pod \"redhat-marketplace-mpzr9\" (UID: \"711fb3fa-94e5-499d-b1f5-ca2fe7326310\") " pod="openshift-marketplace/redhat-marketplace-mpzr9" Dec 03 17:28:09 crc kubenswrapper[4998]: I1203 17:28:09.595141 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/711fb3fa-94e5-499d-b1f5-ca2fe7326310-catalog-content\") pod \"redhat-marketplace-mpzr9\" (UID: \"711fb3fa-94e5-499d-b1f5-ca2fe7326310\") " pod="openshift-marketplace/redhat-marketplace-mpzr9" Dec 03 17:28:09 crc kubenswrapper[4998]: I1203 17:28:09.595304 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/711fb3fa-94e5-499d-b1f5-ca2fe7326310-utilities\") pod \"redhat-marketplace-mpzr9\" (UID: \"711fb3fa-94e5-499d-b1f5-ca2fe7326310\") " pod="openshift-marketplace/redhat-marketplace-mpzr9" Dec 03 17:28:09 crc kubenswrapper[4998]: I1203 17:28:09.610064 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mpzr9"] Dec 03 17:28:09 crc kubenswrapper[4998]: I1203 17:28:09.697228 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/711fb3fa-94e5-499d-b1f5-ca2fe7326310-catalog-content\") pod \"redhat-marketplace-mpzr9\" (UID: \"711fb3fa-94e5-499d-b1f5-ca2fe7326310\") " pod="openshift-marketplace/redhat-marketplace-mpzr9" Dec 03 17:28:09 crc kubenswrapper[4998]: I1203 17:28:09.697412 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/711fb3fa-94e5-499d-b1f5-ca2fe7326310-utilities\") pod \"redhat-marketplace-mpzr9\" (UID: \"711fb3fa-94e5-499d-b1f5-ca2fe7326310\") " pod="openshift-marketplace/redhat-marketplace-mpzr9" Dec 03 17:28:09 crc kubenswrapper[4998]: I1203 17:28:09.697547 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xblmg\" (UniqueName: \"kubernetes.io/projected/711fb3fa-94e5-499d-b1f5-ca2fe7326310-kube-api-access-xblmg\") pod \"redhat-marketplace-mpzr9\" (UID: \"711fb3fa-94e5-499d-b1f5-ca2fe7326310\") " pod="openshift-marketplace/redhat-marketplace-mpzr9" Dec 03 17:28:09 crc kubenswrapper[4998]: I1203 17:28:09.697841 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/711fb3fa-94e5-499d-b1f5-ca2fe7326310-catalog-content\") pod \"redhat-marketplace-mpzr9\" (UID: \"711fb3fa-94e5-499d-b1f5-ca2fe7326310\") " pod="openshift-marketplace/redhat-marketplace-mpzr9" Dec 03 17:28:09 crc kubenswrapper[4998]: I1203 17:28:09.697868 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/711fb3fa-94e5-499d-b1f5-ca2fe7326310-utilities\") pod \"redhat-marketplace-mpzr9\" (UID: \"711fb3fa-94e5-499d-b1f5-ca2fe7326310\") " pod="openshift-marketplace/redhat-marketplace-mpzr9" Dec 03 17:28:09 crc kubenswrapper[4998]: I1203 17:28:09.723428 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xblmg\" (UniqueName: \"kubernetes.io/projected/711fb3fa-94e5-499d-b1f5-ca2fe7326310-kube-api-access-xblmg\") pod \"redhat-marketplace-mpzr9\" (UID: \"711fb3fa-94e5-499d-b1f5-ca2fe7326310\") " pod="openshift-marketplace/redhat-marketplace-mpzr9" Dec 03 17:28:09 crc kubenswrapper[4998]: I1203 17:28:09.927861 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mpzr9" Dec 03 17:28:10 crc kubenswrapper[4998]: I1203 17:28:10.470398 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mpzr9"] Dec 03 17:28:10 crc kubenswrapper[4998]: I1203 17:28:10.677948 4998 scope.go:117] "RemoveContainer" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" Dec 03 17:28:10 crc kubenswrapper[4998]: E1203 17:28:10.678449 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:28:11 crc kubenswrapper[4998]: I1203 17:28:11.293533 4998 generic.go:334] "Generic (PLEG): container finished" podID="711fb3fa-94e5-499d-b1f5-ca2fe7326310" containerID="387e13943ba167c6d94d91eaaf6ef1f48f52739c545522cfaa03b80ac22a578f" exitCode=0 Dec 03 17:28:11 crc kubenswrapper[4998]: I1203 17:28:11.293590 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpzr9" event={"ID":"711fb3fa-94e5-499d-b1f5-ca2fe7326310","Type":"ContainerDied","Data":"387e13943ba167c6d94d91eaaf6ef1f48f52739c545522cfaa03b80ac22a578f"} Dec 03 17:28:11 crc kubenswrapper[4998]: I1203 17:28:11.293835 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpzr9" event={"ID":"711fb3fa-94e5-499d-b1f5-ca2fe7326310","Type":"ContainerStarted","Data":"d7328c962ce72f7cf108b6db8a072e09d4fefd293c962133d8e3fd691dab40ad"} Dec 03 17:28:12 crc kubenswrapper[4998]: I1203 17:28:12.306181 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpzr9" event={"ID":"711fb3fa-94e5-499d-b1f5-ca2fe7326310","Type":"ContainerStarted","Data":"c79e987394dbd3a522a88a9c54c4b7cf03d24d9900fc2f038c9dcd4377ba2928"} Dec 03 17:28:13 crc kubenswrapper[4998]: I1203 17:28:13.317582 4998 generic.go:334] "Generic (PLEG): container finished" podID="711fb3fa-94e5-499d-b1f5-ca2fe7326310" containerID="c79e987394dbd3a522a88a9c54c4b7cf03d24d9900fc2f038c9dcd4377ba2928" exitCode=0 Dec 03 17:28:13 crc kubenswrapper[4998]: I1203 17:28:13.317660 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpzr9" event={"ID":"711fb3fa-94e5-499d-b1f5-ca2fe7326310","Type":"ContainerDied","Data":"c79e987394dbd3a522a88a9c54c4b7cf03d24d9900fc2f038c9dcd4377ba2928"} Dec 03 17:28:14 crc kubenswrapper[4998]: I1203 17:28:14.330647 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpzr9" event={"ID":"711fb3fa-94e5-499d-b1f5-ca2fe7326310","Type":"ContainerStarted","Data":"a2d35805c707d0277080cda25499974687d76b827a296818f2b90b8dec84064d"} Dec 03 17:28:14 crc kubenswrapper[4998]: I1203 17:28:14.355045 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mpzr9" podStartSLOduration=3.743753335 podStartE2EDuration="5.355025737s" podCreationTimestamp="2025-12-03 17:28:09 +0000 UTC" firstStartedPulling="2025-12-03 17:28:11.295527814 +0000 UTC m=+5069.907228027" lastFinishedPulling="2025-12-03 17:28:12.906800206 +0000 UTC m=+5071.518500429" observedRunningTime="2025-12-03 17:28:14.349495989 +0000 UTC m=+5072.961196222" watchObservedRunningTime="2025-12-03 17:28:14.355025737 +0000 UTC m=+5072.966725960" Dec 03 17:28:19 crc kubenswrapper[4998]: I1203 17:28:19.927949 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mpzr9" Dec 03 17:28:19 crc kubenswrapper[4998]: I1203 17:28:19.928729 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mpzr9" Dec 03 17:28:20 crc kubenswrapper[4998]: I1203 17:28:20.015861 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mpzr9" Dec 03 17:28:20 crc kubenswrapper[4998]: I1203 17:28:20.477585 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mpzr9" Dec 03 17:28:20 crc kubenswrapper[4998]: I1203 17:28:20.550516 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mpzr9"] Dec 03 17:28:22 crc kubenswrapper[4998]: I1203 17:28:22.429722 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mpzr9" podUID="711fb3fa-94e5-499d-b1f5-ca2fe7326310" containerName="registry-server" containerID="cri-o://a2d35805c707d0277080cda25499974687d76b827a296818f2b90b8dec84064d" gracePeriod=2 Dec 03 17:28:23 crc kubenswrapper[4998]: I1203 17:28:23.440683 4998 generic.go:334] "Generic (PLEG): container finished" podID="711fb3fa-94e5-499d-b1f5-ca2fe7326310" containerID="a2d35805c707d0277080cda25499974687d76b827a296818f2b90b8dec84064d" exitCode=0 Dec 03 17:28:23 crc kubenswrapper[4998]: I1203 17:28:23.441450 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpzr9" event={"ID":"711fb3fa-94e5-499d-b1f5-ca2fe7326310","Type":"ContainerDied","Data":"a2d35805c707d0277080cda25499974687d76b827a296818f2b90b8dec84064d"} Dec 03 17:28:23 crc kubenswrapper[4998]: I1203 17:28:23.626499 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mpzr9" Dec 03 17:28:23 crc kubenswrapper[4998]: I1203 17:28:23.677807 4998 scope.go:117] "RemoveContainer" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" Dec 03 17:28:23 crc kubenswrapper[4998]: E1203 17:28:23.678148 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:28:23 crc kubenswrapper[4998]: I1203 17:28:23.775667 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xblmg\" (UniqueName: \"kubernetes.io/projected/711fb3fa-94e5-499d-b1f5-ca2fe7326310-kube-api-access-xblmg\") pod \"711fb3fa-94e5-499d-b1f5-ca2fe7326310\" (UID: \"711fb3fa-94e5-499d-b1f5-ca2fe7326310\") " Dec 03 17:28:23 crc kubenswrapper[4998]: I1203 17:28:23.775816 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/711fb3fa-94e5-499d-b1f5-ca2fe7326310-catalog-content\") pod \"711fb3fa-94e5-499d-b1f5-ca2fe7326310\" (UID: \"711fb3fa-94e5-499d-b1f5-ca2fe7326310\") " Dec 03 17:28:23 crc kubenswrapper[4998]: I1203 17:28:23.775870 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/711fb3fa-94e5-499d-b1f5-ca2fe7326310-utilities\") pod \"711fb3fa-94e5-499d-b1f5-ca2fe7326310\" (UID: \"711fb3fa-94e5-499d-b1f5-ca2fe7326310\") " Dec 03 17:28:23 crc kubenswrapper[4998]: I1203 17:28:23.776830 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/711fb3fa-94e5-499d-b1f5-ca2fe7326310-utilities" (OuterVolumeSpecName: "utilities") pod "711fb3fa-94e5-499d-b1f5-ca2fe7326310" (UID: "711fb3fa-94e5-499d-b1f5-ca2fe7326310"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:28:23 crc kubenswrapper[4998]: I1203 17:28:23.785256 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/711fb3fa-94e5-499d-b1f5-ca2fe7326310-kube-api-access-xblmg" (OuterVolumeSpecName: "kube-api-access-xblmg") pod "711fb3fa-94e5-499d-b1f5-ca2fe7326310" (UID: "711fb3fa-94e5-499d-b1f5-ca2fe7326310"). InnerVolumeSpecName "kube-api-access-xblmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:28:23 crc kubenswrapper[4998]: I1203 17:28:23.809436 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/711fb3fa-94e5-499d-b1f5-ca2fe7326310-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "711fb3fa-94e5-499d-b1f5-ca2fe7326310" (UID: "711fb3fa-94e5-499d-b1f5-ca2fe7326310"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:28:23 crc kubenswrapper[4998]: I1203 17:28:23.879534 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xblmg\" (UniqueName: \"kubernetes.io/projected/711fb3fa-94e5-499d-b1f5-ca2fe7326310-kube-api-access-xblmg\") on node \"crc\" DevicePath \"\"" Dec 03 17:28:23 crc kubenswrapper[4998]: I1203 17:28:23.879594 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/711fb3fa-94e5-499d-b1f5-ca2fe7326310-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:28:23 crc kubenswrapper[4998]: I1203 17:28:23.879613 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/711fb3fa-94e5-499d-b1f5-ca2fe7326310-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:28:24 crc kubenswrapper[4998]: I1203 17:28:24.461347 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpzr9" event={"ID":"711fb3fa-94e5-499d-b1f5-ca2fe7326310","Type":"ContainerDied","Data":"d7328c962ce72f7cf108b6db8a072e09d4fefd293c962133d8e3fd691dab40ad"} Dec 03 17:28:24 crc kubenswrapper[4998]: I1203 17:28:24.461453 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mpzr9" Dec 03 17:28:24 crc kubenswrapper[4998]: I1203 17:28:24.463047 4998 scope.go:117] "RemoveContainer" containerID="a2d35805c707d0277080cda25499974687d76b827a296818f2b90b8dec84064d" Dec 03 17:28:24 crc kubenswrapper[4998]: I1203 17:28:24.500825 4998 scope.go:117] "RemoveContainer" containerID="c79e987394dbd3a522a88a9c54c4b7cf03d24d9900fc2f038c9dcd4377ba2928" Dec 03 17:28:24 crc kubenswrapper[4998]: I1203 17:28:24.536048 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mpzr9"] Dec 03 17:28:24 crc kubenswrapper[4998]: I1203 17:28:24.551644 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mpzr9"] Dec 03 17:28:24 crc kubenswrapper[4998]: I1203 17:28:24.567901 4998 scope.go:117] "RemoveContainer" containerID="387e13943ba167c6d94d91eaaf6ef1f48f52739c545522cfaa03b80ac22a578f" Dec 03 17:28:25 crc kubenswrapper[4998]: I1203 17:28:25.702042 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="711fb3fa-94e5-499d-b1f5-ca2fe7326310" path="/var/lib/kubelet/pods/711fb3fa-94e5-499d-b1f5-ca2fe7326310/volumes" Dec 03 17:28:36 crc kubenswrapper[4998]: I1203 17:28:36.677630 4998 scope.go:117] "RemoveContainer" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" Dec 03 17:28:36 crc kubenswrapper[4998]: E1203 17:28:36.678910 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:28:51 crc kubenswrapper[4998]: I1203 17:28:51.690297 4998 scope.go:117] "RemoveContainer" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" Dec 03 17:28:51 crc kubenswrapper[4998]: E1203 17:28:51.692841 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:29:02 crc kubenswrapper[4998]: I1203 17:29:02.677504 4998 scope.go:117] "RemoveContainer" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" Dec 03 17:29:02 crc kubenswrapper[4998]: E1203 17:29:02.678277 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:29:14 crc kubenswrapper[4998]: I1203 17:29:14.678801 4998 scope.go:117] "RemoveContainer" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" Dec 03 17:29:14 crc kubenswrapper[4998]: E1203 17:29:14.680544 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:29:26 crc kubenswrapper[4998]: I1203 17:29:26.963811 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xcjts"] Dec 03 17:29:26 crc kubenswrapper[4998]: E1203 17:29:26.964798 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="711fb3fa-94e5-499d-b1f5-ca2fe7326310" containerName="extract-utilities" Dec 03 17:29:26 crc kubenswrapper[4998]: I1203 17:29:26.964814 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="711fb3fa-94e5-499d-b1f5-ca2fe7326310" containerName="extract-utilities" Dec 03 17:29:26 crc kubenswrapper[4998]: E1203 17:29:26.964840 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="711fb3fa-94e5-499d-b1f5-ca2fe7326310" containerName="registry-server" Dec 03 17:29:26 crc kubenswrapper[4998]: I1203 17:29:26.964845 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="711fb3fa-94e5-499d-b1f5-ca2fe7326310" containerName="registry-server" Dec 03 17:29:26 crc kubenswrapper[4998]: E1203 17:29:26.964864 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="711fb3fa-94e5-499d-b1f5-ca2fe7326310" containerName="extract-content" Dec 03 17:29:26 crc kubenswrapper[4998]: I1203 17:29:26.964872 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="711fb3fa-94e5-499d-b1f5-ca2fe7326310" containerName="extract-content" Dec 03 17:29:26 crc kubenswrapper[4998]: I1203 17:29:26.965078 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="711fb3fa-94e5-499d-b1f5-ca2fe7326310" containerName="registry-server" Dec 03 17:29:26 crc kubenswrapper[4998]: I1203 17:29:26.967523 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xcjts" Dec 03 17:29:26 crc kubenswrapper[4998]: I1203 17:29:26.987151 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xcjts"] Dec 03 17:29:27 crc kubenswrapper[4998]: I1203 17:29:27.160785 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24d16d6a-92b7-4903-90bb-e5bd74988ca4-utilities\") pod \"certified-operators-xcjts\" (UID: \"24d16d6a-92b7-4903-90bb-e5bd74988ca4\") " pod="openshift-marketplace/certified-operators-xcjts" Dec 03 17:29:27 crc kubenswrapper[4998]: I1203 17:29:27.160924 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t85d\" (UniqueName: \"kubernetes.io/projected/24d16d6a-92b7-4903-90bb-e5bd74988ca4-kube-api-access-6t85d\") pod \"certified-operators-xcjts\" (UID: \"24d16d6a-92b7-4903-90bb-e5bd74988ca4\") " pod="openshift-marketplace/certified-operators-xcjts" Dec 03 17:29:27 crc kubenswrapper[4998]: I1203 17:29:27.161870 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24d16d6a-92b7-4903-90bb-e5bd74988ca4-catalog-content\") pod \"certified-operators-xcjts\" (UID: \"24d16d6a-92b7-4903-90bb-e5bd74988ca4\") " pod="openshift-marketplace/certified-operators-xcjts" Dec 03 17:29:27 crc kubenswrapper[4998]: I1203 17:29:27.264373 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t85d\" (UniqueName: \"kubernetes.io/projected/24d16d6a-92b7-4903-90bb-e5bd74988ca4-kube-api-access-6t85d\") pod \"certified-operators-xcjts\" (UID: \"24d16d6a-92b7-4903-90bb-e5bd74988ca4\") " pod="openshift-marketplace/certified-operators-xcjts" Dec 03 17:29:27 crc kubenswrapper[4998]: I1203 17:29:27.264507 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24d16d6a-92b7-4903-90bb-e5bd74988ca4-catalog-content\") pod \"certified-operators-xcjts\" (UID: \"24d16d6a-92b7-4903-90bb-e5bd74988ca4\") " pod="openshift-marketplace/certified-operators-xcjts" Dec 03 17:29:27 crc kubenswrapper[4998]: I1203 17:29:27.264641 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24d16d6a-92b7-4903-90bb-e5bd74988ca4-utilities\") pod \"certified-operators-xcjts\" (UID: \"24d16d6a-92b7-4903-90bb-e5bd74988ca4\") " pod="openshift-marketplace/certified-operators-xcjts" Dec 03 17:29:27 crc kubenswrapper[4998]: I1203 17:29:27.265158 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24d16d6a-92b7-4903-90bb-e5bd74988ca4-catalog-content\") pod \"certified-operators-xcjts\" (UID: \"24d16d6a-92b7-4903-90bb-e5bd74988ca4\") " pod="openshift-marketplace/certified-operators-xcjts" Dec 03 17:29:27 crc kubenswrapper[4998]: I1203 17:29:27.265157 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24d16d6a-92b7-4903-90bb-e5bd74988ca4-utilities\") pod \"certified-operators-xcjts\" (UID: \"24d16d6a-92b7-4903-90bb-e5bd74988ca4\") " pod="openshift-marketplace/certified-operators-xcjts" Dec 03 17:29:27 crc kubenswrapper[4998]: I1203 17:29:27.789837 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t85d\" (UniqueName: \"kubernetes.io/projected/24d16d6a-92b7-4903-90bb-e5bd74988ca4-kube-api-access-6t85d\") pod \"certified-operators-xcjts\" (UID: \"24d16d6a-92b7-4903-90bb-e5bd74988ca4\") " pod="openshift-marketplace/certified-operators-xcjts" Dec 03 17:29:27 crc kubenswrapper[4998]: I1203 17:29:27.909154 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xcjts" Dec 03 17:29:28 crc kubenswrapper[4998]: I1203 17:29:28.522535 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xcjts"] Dec 03 17:29:28 crc kubenswrapper[4998]: I1203 17:29:28.678363 4998 scope.go:117] "RemoveContainer" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" Dec 03 17:29:28 crc kubenswrapper[4998]: E1203 17:29:28.679231 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:29:29 crc kubenswrapper[4998]: I1203 17:29:29.155984 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fbz7l"] Dec 03 17:29:29 crc kubenswrapper[4998]: I1203 17:29:29.158660 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fbz7l" Dec 03 17:29:29 crc kubenswrapper[4998]: I1203 17:29:29.173442 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fbz7l"] Dec 03 17:29:29 crc kubenswrapper[4998]: I1203 17:29:29.252474 4998 generic.go:334] "Generic (PLEG): container finished" podID="24d16d6a-92b7-4903-90bb-e5bd74988ca4" containerID="0afcafe6ba6c13135aa03f25e261702424c6df3c4c1c03c68e43edf95b0a0e97" exitCode=0 Dec 03 17:29:29 crc kubenswrapper[4998]: I1203 17:29:29.252525 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xcjts" event={"ID":"24d16d6a-92b7-4903-90bb-e5bd74988ca4","Type":"ContainerDied","Data":"0afcafe6ba6c13135aa03f25e261702424c6df3c4c1c03c68e43edf95b0a0e97"} Dec 03 17:29:29 crc kubenswrapper[4998]: I1203 17:29:29.252561 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xcjts" event={"ID":"24d16d6a-92b7-4903-90bb-e5bd74988ca4","Type":"ContainerStarted","Data":"68a467e940014030963e4a24df4f9af7edd3a9db7821f1cb9cb3e5ae20b53898"} Dec 03 17:29:29 crc kubenswrapper[4998]: I1203 17:29:29.325720 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb0309b8-299a-47f9-9e4b-57ed719655ab-utilities\") pod \"community-operators-fbz7l\" (UID: \"fb0309b8-299a-47f9-9e4b-57ed719655ab\") " pod="openshift-marketplace/community-operators-fbz7l" Dec 03 17:29:29 crc kubenswrapper[4998]: I1203 17:29:29.326144 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68snf\" (UniqueName: \"kubernetes.io/projected/fb0309b8-299a-47f9-9e4b-57ed719655ab-kube-api-access-68snf\") pod \"community-operators-fbz7l\" (UID: \"fb0309b8-299a-47f9-9e4b-57ed719655ab\") " pod="openshift-marketplace/community-operators-fbz7l" Dec 03 17:29:29 crc kubenswrapper[4998]: I1203 17:29:29.326247 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb0309b8-299a-47f9-9e4b-57ed719655ab-catalog-content\") pod \"community-operators-fbz7l\" (UID: \"fb0309b8-299a-47f9-9e4b-57ed719655ab\") " pod="openshift-marketplace/community-operators-fbz7l" Dec 03 17:29:29 crc kubenswrapper[4998]: I1203 17:29:29.428837 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb0309b8-299a-47f9-9e4b-57ed719655ab-utilities\") pod \"community-operators-fbz7l\" (UID: \"fb0309b8-299a-47f9-9e4b-57ed719655ab\") " pod="openshift-marketplace/community-operators-fbz7l" Dec 03 17:29:29 crc kubenswrapper[4998]: I1203 17:29:29.428971 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68snf\" (UniqueName: \"kubernetes.io/projected/fb0309b8-299a-47f9-9e4b-57ed719655ab-kube-api-access-68snf\") pod \"community-operators-fbz7l\" (UID: \"fb0309b8-299a-47f9-9e4b-57ed719655ab\") " pod="openshift-marketplace/community-operators-fbz7l" Dec 03 17:29:29 crc kubenswrapper[4998]: I1203 17:29:29.429083 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb0309b8-299a-47f9-9e4b-57ed719655ab-catalog-content\") pod \"community-operators-fbz7l\" (UID: \"fb0309b8-299a-47f9-9e4b-57ed719655ab\") " pod="openshift-marketplace/community-operators-fbz7l" Dec 03 17:29:29 crc kubenswrapper[4998]: I1203 17:29:29.429349 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb0309b8-299a-47f9-9e4b-57ed719655ab-utilities\") pod \"community-operators-fbz7l\" (UID: \"fb0309b8-299a-47f9-9e4b-57ed719655ab\") " pod="openshift-marketplace/community-operators-fbz7l" Dec 03 17:29:29 crc kubenswrapper[4998]: I1203 17:29:29.429673 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb0309b8-299a-47f9-9e4b-57ed719655ab-catalog-content\") pod \"community-operators-fbz7l\" (UID: \"fb0309b8-299a-47f9-9e4b-57ed719655ab\") " pod="openshift-marketplace/community-operators-fbz7l" Dec 03 17:29:29 crc kubenswrapper[4998]: I1203 17:29:29.695510 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68snf\" (UniqueName: \"kubernetes.io/projected/fb0309b8-299a-47f9-9e4b-57ed719655ab-kube-api-access-68snf\") pod \"community-operators-fbz7l\" (UID: \"fb0309b8-299a-47f9-9e4b-57ed719655ab\") " pod="openshift-marketplace/community-operators-fbz7l" Dec 03 17:29:29 crc kubenswrapper[4998]: I1203 17:29:29.850821 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fbz7l" Dec 03 17:29:30 crc kubenswrapper[4998]: W1203 17:29:30.390897 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb0309b8_299a_47f9_9e4b_57ed719655ab.slice/crio-c1dd706676c6532f42763d06ab24e5e538719abe4b6cfac219e4bea5ac8f5527 WatchSource:0}: Error finding container c1dd706676c6532f42763d06ab24e5e538719abe4b6cfac219e4bea5ac8f5527: Status 404 returned error can't find the container with id c1dd706676c6532f42763d06ab24e5e538719abe4b6cfac219e4bea5ac8f5527 Dec 03 17:29:30 crc kubenswrapper[4998]: I1203 17:29:30.412046 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fbz7l"] Dec 03 17:29:31 crc kubenswrapper[4998]: I1203 17:29:31.275245 4998 generic.go:334] "Generic (PLEG): container finished" podID="fb0309b8-299a-47f9-9e4b-57ed719655ab" containerID="cde15624f81b7f6a9224b3abb7c1baacedfc0b9559dd9b989f5b131a491075c6" exitCode=0 Dec 03 17:29:31 crc kubenswrapper[4998]: I1203 17:29:31.275309 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fbz7l" event={"ID":"fb0309b8-299a-47f9-9e4b-57ed719655ab","Type":"ContainerDied","Data":"cde15624f81b7f6a9224b3abb7c1baacedfc0b9559dd9b989f5b131a491075c6"} Dec 03 17:29:31 crc kubenswrapper[4998]: I1203 17:29:31.275627 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fbz7l" event={"ID":"fb0309b8-299a-47f9-9e4b-57ed719655ab","Type":"ContainerStarted","Data":"c1dd706676c6532f42763d06ab24e5e538719abe4b6cfac219e4bea5ac8f5527"} Dec 03 17:29:31 crc kubenswrapper[4998]: I1203 17:29:31.279858 4998 generic.go:334] "Generic (PLEG): container finished" podID="24d16d6a-92b7-4903-90bb-e5bd74988ca4" containerID="e42b1dfe02a59f90c3f268e9a2c18e2ae9c0af1b9975dcd8fdb634eb5f851389" exitCode=0 Dec 03 17:29:31 crc kubenswrapper[4998]: I1203 17:29:31.279892 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xcjts" event={"ID":"24d16d6a-92b7-4903-90bb-e5bd74988ca4","Type":"ContainerDied","Data":"e42b1dfe02a59f90c3f268e9a2c18e2ae9c0af1b9975dcd8fdb634eb5f851389"} Dec 03 17:29:32 crc kubenswrapper[4998]: I1203 17:29:32.292077 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fbz7l" event={"ID":"fb0309b8-299a-47f9-9e4b-57ed719655ab","Type":"ContainerStarted","Data":"b9f33af3bda317c03b4f2dc0f5b6ddeff6c38574f0724a9793c79c9c7e5b8cb2"} Dec 03 17:29:32 crc kubenswrapper[4998]: I1203 17:29:32.295680 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xcjts" event={"ID":"24d16d6a-92b7-4903-90bb-e5bd74988ca4","Type":"ContainerStarted","Data":"353cefa5e8dcbd9fa3216b2a8a4467d9398681e892ad9609487f7ec1445008cc"} Dec 03 17:29:32 crc kubenswrapper[4998]: I1203 17:29:32.356463 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xcjts" podStartSLOduration=3.825618834 podStartE2EDuration="6.356440491s" podCreationTimestamp="2025-12-03 17:29:26 +0000 UTC" firstStartedPulling="2025-12-03 17:29:29.25569182 +0000 UTC m=+5147.867392043" lastFinishedPulling="2025-12-03 17:29:31.786513447 +0000 UTC m=+5150.398213700" observedRunningTime="2025-12-03 17:29:32.34677828 +0000 UTC m=+5150.958478523" watchObservedRunningTime="2025-12-03 17:29:32.356440491 +0000 UTC m=+5150.968140714" Dec 03 17:29:34 crc kubenswrapper[4998]: I1203 17:29:34.325094 4998 generic.go:334] "Generic (PLEG): container finished" podID="fb0309b8-299a-47f9-9e4b-57ed719655ab" containerID="b9f33af3bda317c03b4f2dc0f5b6ddeff6c38574f0724a9793c79c9c7e5b8cb2" exitCode=0 Dec 03 17:29:34 crc kubenswrapper[4998]: I1203 17:29:34.325188 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fbz7l" event={"ID":"fb0309b8-299a-47f9-9e4b-57ed719655ab","Type":"ContainerDied","Data":"b9f33af3bda317c03b4f2dc0f5b6ddeff6c38574f0724a9793c79c9c7e5b8cb2"} Dec 03 17:29:34 crc kubenswrapper[4998]: I1203 17:29:34.328205 4998 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 17:29:35 crc kubenswrapper[4998]: I1203 17:29:35.341814 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fbz7l" event={"ID":"fb0309b8-299a-47f9-9e4b-57ed719655ab","Type":"ContainerStarted","Data":"4518a4fb8779ec119690ae0a8957da0adf8e989881b391960c15e3354205c017"} Dec 03 17:29:35 crc kubenswrapper[4998]: I1203 17:29:35.383491 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fbz7l" podStartSLOduration=2.891599355 podStartE2EDuration="6.383466817s" podCreationTimestamp="2025-12-03 17:29:29 +0000 UTC" firstStartedPulling="2025-12-03 17:29:31.277347605 +0000 UTC m=+5149.889047828" lastFinishedPulling="2025-12-03 17:29:34.769215037 +0000 UTC m=+5153.380915290" observedRunningTime="2025-12-03 17:29:35.366613667 +0000 UTC m=+5153.978313920" watchObservedRunningTime="2025-12-03 17:29:35.383466817 +0000 UTC m=+5153.995167060" Dec 03 17:29:37 crc kubenswrapper[4998]: I1203 17:29:37.910325 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xcjts" Dec 03 17:29:37 crc kubenswrapper[4998]: I1203 17:29:37.911974 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xcjts" Dec 03 17:29:38 crc kubenswrapper[4998]: I1203 17:29:38.004842 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xcjts" Dec 03 17:29:39 crc kubenswrapper[4998]: I1203 17:29:39.146712 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xcjts" Dec 03 17:29:39 crc kubenswrapper[4998]: I1203 17:29:39.853810 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fbz7l" Dec 03 17:29:39 crc kubenswrapper[4998]: I1203 17:29:39.854205 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fbz7l" Dec 03 17:29:39 crc kubenswrapper[4998]: I1203 17:29:39.911273 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fbz7l" Dec 03 17:29:40 crc kubenswrapper[4998]: I1203 17:29:40.344995 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xcjts"] Dec 03 17:29:40 crc kubenswrapper[4998]: I1203 17:29:40.407705 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xcjts" podUID="24d16d6a-92b7-4903-90bb-e5bd74988ca4" containerName="registry-server" containerID="cri-o://353cefa5e8dcbd9fa3216b2a8a4467d9398681e892ad9609487f7ec1445008cc" gracePeriod=2 Dec 03 17:29:40 crc kubenswrapper[4998]: I1203 17:29:40.458155 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fbz7l" Dec 03 17:29:41 crc kubenswrapper[4998]: I1203 17:29:41.420941 4998 generic.go:334] "Generic (PLEG): container finished" podID="24d16d6a-92b7-4903-90bb-e5bd74988ca4" containerID="353cefa5e8dcbd9fa3216b2a8a4467d9398681e892ad9609487f7ec1445008cc" exitCode=0 Dec 03 17:29:41 crc kubenswrapper[4998]: I1203 17:29:41.421011 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xcjts" event={"ID":"24d16d6a-92b7-4903-90bb-e5bd74988ca4","Type":"ContainerDied","Data":"353cefa5e8dcbd9fa3216b2a8a4467d9398681e892ad9609487f7ec1445008cc"} Dec 03 17:29:42 crc kubenswrapper[4998]: I1203 17:29:42.133101 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xcjts" Dec 03 17:29:42 crc kubenswrapper[4998]: I1203 17:29:42.145178 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fbz7l"] Dec 03 17:29:42 crc kubenswrapper[4998]: I1203 17:29:42.301496 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24d16d6a-92b7-4903-90bb-e5bd74988ca4-catalog-content\") pod \"24d16d6a-92b7-4903-90bb-e5bd74988ca4\" (UID: \"24d16d6a-92b7-4903-90bb-e5bd74988ca4\") " Dec 03 17:29:42 crc kubenswrapper[4998]: I1203 17:29:42.301686 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24d16d6a-92b7-4903-90bb-e5bd74988ca4-utilities\") pod \"24d16d6a-92b7-4903-90bb-e5bd74988ca4\" (UID: \"24d16d6a-92b7-4903-90bb-e5bd74988ca4\") " Dec 03 17:29:42 crc kubenswrapper[4998]: I1203 17:29:42.301718 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6t85d\" (UniqueName: \"kubernetes.io/projected/24d16d6a-92b7-4903-90bb-e5bd74988ca4-kube-api-access-6t85d\") pod \"24d16d6a-92b7-4903-90bb-e5bd74988ca4\" (UID: \"24d16d6a-92b7-4903-90bb-e5bd74988ca4\") " Dec 03 17:29:42 crc kubenswrapper[4998]: I1203 17:29:42.302984 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24d16d6a-92b7-4903-90bb-e5bd74988ca4-utilities" (OuterVolumeSpecName: "utilities") pod "24d16d6a-92b7-4903-90bb-e5bd74988ca4" (UID: "24d16d6a-92b7-4903-90bb-e5bd74988ca4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:29:42 crc kubenswrapper[4998]: I1203 17:29:42.308988 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24d16d6a-92b7-4903-90bb-e5bd74988ca4-kube-api-access-6t85d" (OuterVolumeSpecName: "kube-api-access-6t85d") pod "24d16d6a-92b7-4903-90bb-e5bd74988ca4" (UID: "24d16d6a-92b7-4903-90bb-e5bd74988ca4"). InnerVolumeSpecName "kube-api-access-6t85d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:29:42 crc kubenswrapper[4998]: I1203 17:29:42.364291 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24d16d6a-92b7-4903-90bb-e5bd74988ca4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "24d16d6a-92b7-4903-90bb-e5bd74988ca4" (UID: "24d16d6a-92b7-4903-90bb-e5bd74988ca4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:29:42 crc kubenswrapper[4998]: I1203 17:29:42.404166 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24d16d6a-92b7-4903-90bb-e5bd74988ca4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:29:42 crc kubenswrapper[4998]: I1203 17:29:42.404210 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24d16d6a-92b7-4903-90bb-e5bd74988ca4-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:29:42 crc kubenswrapper[4998]: I1203 17:29:42.404220 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6t85d\" (UniqueName: \"kubernetes.io/projected/24d16d6a-92b7-4903-90bb-e5bd74988ca4-kube-api-access-6t85d\") on node \"crc\" DevicePath \"\"" Dec 03 17:29:42 crc kubenswrapper[4998]: I1203 17:29:42.435540 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fbz7l" podUID="fb0309b8-299a-47f9-9e4b-57ed719655ab" containerName="registry-server" containerID="cri-o://4518a4fb8779ec119690ae0a8957da0adf8e989881b391960c15e3354205c017" gracePeriod=2 Dec 03 17:29:42 crc kubenswrapper[4998]: I1203 17:29:42.435937 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xcjts" Dec 03 17:29:42 crc kubenswrapper[4998]: I1203 17:29:42.436059 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xcjts" event={"ID":"24d16d6a-92b7-4903-90bb-e5bd74988ca4","Type":"ContainerDied","Data":"68a467e940014030963e4a24df4f9af7edd3a9db7821f1cb9cb3e5ae20b53898"} Dec 03 17:29:42 crc kubenswrapper[4998]: I1203 17:29:42.436165 4998 scope.go:117] "RemoveContainer" containerID="353cefa5e8dcbd9fa3216b2a8a4467d9398681e892ad9609487f7ec1445008cc" Dec 03 17:29:42 crc kubenswrapper[4998]: I1203 17:29:42.485956 4998 scope.go:117] "RemoveContainer" containerID="e42b1dfe02a59f90c3f268e9a2c18e2ae9c0af1b9975dcd8fdb634eb5f851389" Dec 03 17:29:42 crc kubenswrapper[4998]: I1203 17:29:42.501515 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xcjts"] Dec 03 17:29:42 crc kubenswrapper[4998]: I1203 17:29:42.510268 4998 scope.go:117] "RemoveContainer" containerID="0afcafe6ba6c13135aa03f25e261702424c6df3c4c1c03c68e43edf95b0a0e97" Dec 03 17:29:42 crc kubenswrapper[4998]: I1203 17:29:42.511679 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xcjts"] Dec 03 17:29:42 crc kubenswrapper[4998]: I1203 17:29:42.874644 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fbz7l" Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.018104 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb0309b8-299a-47f9-9e4b-57ed719655ab-utilities\") pod \"fb0309b8-299a-47f9-9e4b-57ed719655ab\" (UID: \"fb0309b8-299a-47f9-9e4b-57ed719655ab\") " Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.018210 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68snf\" (UniqueName: \"kubernetes.io/projected/fb0309b8-299a-47f9-9e4b-57ed719655ab-kube-api-access-68snf\") pod \"fb0309b8-299a-47f9-9e4b-57ed719655ab\" (UID: \"fb0309b8-299a-47f9-9e4b-57ed719655ab\") " Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.018298 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb0309b8-299a-47f9-9e4b-57ed719655ab-catalog-content\") pod \"fb0309b8-299a-47f9-9e4b-57ed719655ab\" (UID: \"fb0309b8-299a-47f9-9e4b-57ed719655ab\") " Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.019016 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb0309b8-299a-47f9-9e4b-57ed719655ab-utilities" (OuterVolumeSpecName: "utilities") pod "fb0309b8-299a-47f9-9e4b-57ed719655ab" (UID: "fb0309b8-299a-47f9-9e4b-57ed719655ab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.023601 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb0309b8-299a-47f9-9e4b-57ed719655ab-kube-api-access-68snf" (OuterVolumeSpecName: "kube-api-access-68snf") pod "fb0309b8-299a-47f9-9e4b-57ed719655ab" (UID: "fb0309b8-299a-47f9-9e4b-57ed719655ab"). InnerVolumeSpecName "kube-api-access-68snf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.065553 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb0309b8-299a-47f9-9e4b-57ed719655ab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fb0309b8-299a-47f9-9e4b-57ed719655ab" (UID: "fb0309b8-299a-47f9-9e4b-57ed719655ab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.121246 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb0309b8-299a-47f9-9e4b-57ed719655ab-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.121302 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68snf\" (UniqueName: \"kubernetes.io/projected/fb0309b8-299a-47f9-9e4b-57ed719655ab-kube-api-access-68snf\") on node \"crc\" DevicePath \"\"" Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.121325 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb0309b8-299a-47f9-9e4b-57ed719655ab-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.448831 4998 generic.go:334] "Generic (PLEG): container finished" podID="fb0309b8-299a-47f9-9e4b-57ed719655ab" containerID="4518a4fb8779ec119690ae0a8957da0adf8e989881b391960c15e3354205c017" exitCode=0 Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.448908 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fbz7l" event={"ID":"fb0309b8-299a-47f9-9e4b-57ed719655ab","Type":"ContainerDied","Data":"4518a4fb8779ec119690ae0a8957da0adf8e989881b391960c15e3354205c017"} Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.448977 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fbz7l" Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.449010 4998 scope.go:117] "RemoveContainer" containerID="4518a4fb8779ec119690ae0a8957da0adf8e989881b391960c15e3354205c017" Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.448985 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fbz7l" event={"ID":"fb0309b8-299a-47f9-9e4b-57ed719655ab","Type":"ContainerDied","Data":"c1dd706676c6532f42763d06ab24e5e538719abe4b6cfac219e4bea5ac8f5527"} Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.479991 4998 scope.go:117] "RemoveContainer" containerID="b9f33af3bda317c03b4f2dc0f5b6ddeff6c38574f0724a9793c79c9c7e5b8cb2" Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.503489 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fbz7l"] Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.516025 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fbz7l"] Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.533028 4998 scope.go:117] "RemoveContainer" containerID="cde15624f81b7f6a9224b3abb7c1baacedfc0b9559dd9b989f5b131a491075c6" Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.576361 4998 scope.go:117] "RemoveContainer" containerID="4518a4fb8779ec119690ae0a8957da0adf8e989881b391960c15e3354205c017" Dec 03 17:29:43 crc kubenswrapper[4998]: E1203 17:29:43.576979 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4518a4fb8779ec119690ae0a8957da0adf8e989881b391960c15e3354205c017\": container with ID starting with 4518a4fb8779ec119690ae0a8957da0adf8e989881b391960c15e3354205c017 not found: ID does not exist" containerID="4518a4fb8779ec119690ae0a8957da0adf8e989881b391960c15e3354205c017" Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.577041 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4518a4fb8779ec119690ae0a8957da0adf8e989881b391960c15e3354205c017"} err="failed to get container status \"4518a4fb8779ec119690ae0a8957da0adf8e989881b391960c15e3354205c017\": rpc error: code = NotFound desc = could not find container \"4518a4fb8779ec119690ae0a8957da0adf8e989881b391960c15e3354205c017\": container with ID starting with 4518a4fb8779ec119690ae0a8957da0adf8e989881b391960c15e3354205c017 not found: ID does not exist" Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.577075 4998 scope.go:117] "RemoveContainer" containerID="b9f33af3bda317c03b4f2dc0f5b6ddeff6c38574f0724a9793c79c9c7e5b8cb2" Dec 03 17:29:43 crc kubenswrapper[4998]: E1203 17:29:43.577472 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9f33af3bda317c03b4f2dc0f5b6ddeff6c38574f0724a9793c79c9c7e5b8cb2\": container with ID starting with b9f33af3bda317c03b4f2dc0f5b6ddeff6c38574f0724a9793c79c9c7e5b8cb2 not found: ID does not exist" containerID="b9f33af3bda317c03b4f2dc0f5b6ddeff6c38574f0724a9793c79c9c7e5b8cb2" Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.577511 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9f33af3bda317c03b4f2dc0f5b6ddeff6c38574f0724a9793c79c9c7e5b8cb2"} err="failed to get container status \"b9f33af3bda317c03b4f2dc0f5b6ddeff6c38574f0724a9793c79c9c7e5b8cb2\": rpc error: code = NotFound desc = could not find container \"b9f33af3bda317c03b4f2dc0f5b6ddeff6c38574f0724a9793c79c9c7e5b8cb2\": container with ID starting with b9f33af3bda317c03b4f2dc0f5b6ddeff6c38574f0724a9793c79c9c7e5b8cb2 not found: ID does not exist" Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.577542 4998 scope.go:117] "RemoveContainer" containerID="cde15624f81b7f6a9224b3abb7c1baacedfc0b9559dd9b989f5b131a491075c6" Dec 03 17:29:43 crc kubenswrapper[4998]: E1203 17:29:43.577992 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cde15624f81b7f6a9224b3abb7c1baacedfc0b9559dd9b989f5b131a491075c6\": container with ID starting with cde15624f81b7f6a9224b3abb7c1baacedfc0b9559dd9b989f5b131a491075c6 not found: ID does not exist" containerID="cde15624f81b7f6a9224b3abb7c1baacedfc0b9559dd9b989f5b131a491075c6" Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.578039 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cde15624f81b7f6a9224b3abb7c1baacedfc0b9559dd9b989f5b131a491075c6"} err="failed to get container status \"cde15624f81b7f6a9224b3abb7c1baacedfc0b9559dd9b989f5b131a491075c6\": rpc error: code = NotFound desc = could not find container \"cde15624f81b7f6a9224b3abb7c1baacedfc0b9559dd9b989f5b131a491075c6\": container with ID starting with cde15624f81b7f6a9224b3abb7c1baacedfc0b9559dd9b989f5b131a491075c6 not found: ID does not exist" Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.678511 4998 scope.go:117] "RemoveContainer" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" Dec 03 17:29:43 crc kubenswrapper[4998]: E1203 17:29:43.679822 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.705980 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24d16d6a-92b7-4903-90bb-e5bd74988ca4" path="/var/lib/kubelet/pods/24d16d6a-92b7-4903-90bb-e5bd74988ca4/volumes" Dec 03 17:29:43 crc kubenswrapper[4998]: I1203 17:29:43.707747 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb0309b8-299a-47f9-9e4b-57ed719655ab" path="/var/lib/kubelet/pods/fb0309b8-299a-47f9-9e4b-57ed719655ab/volumes" Dec 03 17:29:55 crc kubenswrapper[4998]: I1203 17:29:55.678989 4998 scope.go:117] "RemoveContainer" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" Dec 03 17:29:55 crc kubenswrapper[4998]: E1203 17:29:55.680366 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:30:00 crc kubenswrapper[4998]: I1203 17:30:00.153889 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413050-vf5wq"] Dec 03 17:30:00 crc kubenswrapper[4998]: E1203 17:30:00.156166 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb0309b8-299a-47f9-9e4b-57ed719655ab" containerName="extract-utilities" Dec 03 17:30:00 crc kubenswrapper[4998]: I1203 17:30:00.156193 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb0309b8-299a-47f9-9e4b-57ed719655ab" containerName="extract-utilities" Dec 03 17:30:00 crc kubenswrapper[4998]: E1203 17:30:00.156205 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24d16d6a-92b7-4903-90bb-e5bd74988ca4" containerName="registry-server" Dec 03 17:30:00 crc kubenswrapper[4998]: I1203 17:30:00.156215 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="24d16d6a-92b7-4903-90bb-e5bd74988ca4" containerName="registry-server" Dec 03 17:30:00 crc kubenswrapper[4998]: E1203 17:30:00.156235 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24d16d6a-92b7-4903-90bb-e5bd74988ca4" containerName="extract-content" Dec 03 17:30:00 crc kubenswrapper[4998]: I1203 17:30:00.156243 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="24d16d6a-92b7-4903-90bb-e5bd74988ca4" containerName="extract-content" Dec 03 17:30:00 crc kubenswrapper[4998]: E1203 17:30:00.156260 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb0309b8-299a-47f9-9e4b-57ed719655ab" containerName="extract-content" Dec 03 17:30:00 crc kubenswrapper[4998]: I1203 17:30:00.156268 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb0309b8-299a-47f9-9e4b-57ed719655ab" containerName="extract-content" Dec 03 17:30:00 crc kubenswrapper[4998]: E1203 17:30:00.156299 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb0309b8-299a-47f9-9e4b-57ed719655ab" containerName="registry-server" Dec 03 17:30:00 crc kubenswrapper[4998]: I1203 17:30:00.156306 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb0309b8-299a-47f9-9e4b-57ed719655ab" containerName="registry-server" Dec 03 17:30:00 crc kubenswrapper[4998]: E1203 17:30:00.156319 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24d16d6a-92b7-4903-90bb-e5bd74988ca4" containerName="extract-utilities" Dec 03 17:30:00 crc kubenswrapper[4998]: I1203 17:30:00.156325 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="24d16d6a-92b7-4903-90bb-e5bd74988ca4" containerName="extract-utilities" Dec 03 17:30:00 crc kubenswrapper[4998]: I1203 17:30:00.156554 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="24d16d6a-92b7-4903-90bb-e5bd74988ca4" containerName="registry-server" Dec 03 17:30:00 crc kubenswrapper[4998]: I1203 17:30:00.156575 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb0309b8-299a-47f9-9e4b-57ed719655ab" containerName="registry-server" Dec 03 17:30:00 crc kubenswrapper[4998]: I1203 17:30:00.157428 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-vf5wq" Dec 03 17:30:00 crc kubenswrapper[4998]: I1203 17:30:00.161200 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 17:30:00 crc kubenswrapper[4998]: I1203 17:30:00.161233 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 17:30:00 crc kubenswrapper[4998]: I1203 17:30:00.173004 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413050-vf5wq"] Dec 03 17:30:00 crc kubenswrapper[4998]: I1203 17:30:00.320083 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwbvv\" (UniqueName: \"kubernetes.io/projected/96968fa5-44ed-4109-8935-684bc644ff19-kube-api-access-xwbvv\") pod \"collect-profiles-29413050-vf5wq\" (UID: \"96968fa5-44ed-4109-8935-684bc644ff19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-vf5wq" Dec 03 17:30:00 crc kubenswrapper[4998]: I1203 17:30:00.320321 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96968fa5-44ed-4109-8935-684bc644ff19-config-volume\") pod \"collect-profiles-29413050-vf5wq\" (UID: \"96968fa5-44ed-4109-8935-684bc644ff19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-vf5wq" Dec 03 17:30:00 crc kubenswrapper[4998]: I1203 17:30:00.320383 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96968fa5-44ed-4109-8935-684bc644ff19-secret-volume\") pod \"collect-profiles-29413050-vf5wq\" (UID: \"96968fa5-44ed-4109-8935-684bc644ff19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-vf5wq" Dec 03 17:30:00 crc kubenswrapper[4998]: I1203 17:30:00.421703 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwbvv\" (UniqueName: \"kubernetes.io/projected/96968fa5-44ed-4109-8935-684bc644ff19-kube-api-access-xwbvv\") pod \"collect-profiles-29413050-vf5wq\" (UID: \"96968fa5-44ed-4109-8935-684bc644ff19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-vf5wq" Dec 03 17:30:00 crc kubenswrapper[4998]: I1203 17:30:00.422171 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96968fa5-44ed-4109-8935-684bc644ff19-config-volume\") pod \"collect-profiles-29413050-vf5wq\" (UID: \"96968fa5-44ed-4109-8935-684bc644ff19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-vf5wq" Dec 03 17:30:00 crc kubenswrapper[4998]: I1203 17:30:00.422230 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96968fa5-44ed-4109-8935-684bc644ff19-secret-volume\") pod \"collect-profiles-29413050-vf5wq\" (UID: \"96968fa5-44ed-4109-8935-684bc644ff19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-vf5wq" Dec 03 17:30:00 crc kubenswrapper[4998]: I1203 17:30:00.423555 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96968fa5-44ed-4109-8935-684bc644ff19-config-volume\") pod \"collect-profiles-29413050-vf5wq\" (UID: \"96968fa5-44ed-4109-8935-684bc644ff19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-vf5wq" Dec 03 17:30:00 crc kubenswrapper[4998]: I1203 17:30:00.429513 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96968fa5-44ed-4109-8935-684bc644ff19-secret-volume\") pod \"collect-profiles-29413050-vf5wq\" (UID: \"96968fa5-44ed-4109-8935-684bc644ff19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-vf5wq" Dec 03 17:30:00 crc kubenswrapper[4998]: I1203 17:30:00.440211 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwbvv\" (UniqueName: \"kubernetes.io/projected/96968fa5-44ed-4109-8935-684bc644ff19-kube-api-access-xwbvv\") pod \"collect-profiles-29413050-vf5wq\" (UID: \"96968fa5-44ed-4109-8935-684bc644ff19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-vf5wq" Dec 03 17:30:00 crc kubenswrapper[4998]: I1203 17:30:00.500153 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-vf5wq" Dec 03 17:30:00 crc kubenswrapper[4998]: I1203 17:30:00.941647 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413050-vf5wq"] Dec 03 17:30:00 crc kubenswrapper[4998]: W1203 17:30:00.947602 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96968fa5_44ed_4109_8935_684bc644ff19.slice/crio-e01f65ec7243d5e8d4d4bbe242639ed5266f6ec4a258b83ab17357a1be414058 WatchSource:0}: Error finding container e01f65ec7243d5e8d4d4bbe242639ed5266f6ec4a258b83ab17357a1be414058: Status 404 returned error can't find the container with id e01f65ec7243d5e8d4d4bbe242639ed5266f6ec4a258b83ab17357a1be414058 Dec 03 17:30:01 crc kubenswrapper[4998]: I1203 17:30:01.686950 4998 generic.go:334] "Generic (PLEG): container finished" podID="96968fa5-44ed-4109-8935-684bc644ff19" containerID="3eed5d77d0099900c09003423c97beb5512674b478a462f388cb6775aeaa647f" exitCode=0 Dec 03 17:30:01 crc kubenswrapper[4998]: I1203 17:30:01.698708 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-vf5wq" event={"ID":"96968fa5-44ed-4109-8935-684bc644ff19","Type":"ContainerDied","Data":"3eed5d77d0099900c09003423c97beb5512674b478a462f388cb6775aeaa647f"} Dec 03 17:30:01 crc kubenswrapper[4998]: I1203 17:30:01.698863 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-vf5wq" event={"ID":"96968fa5-44ed-4109-8935-684bc644ff19","Type":"ContainerStarted","Data":"e01f65ec7243d5e8d4d4bbe242639ed5266f6ec4a258b83ab17357a1be414058"} Dec 03 17:30:03 crc kubenswrapper[4998]: I1203 17:30:03.136927 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-vf5wq" Dec 03 17:30:03 crc kubenswrapper[4998]: I1203 17:30:03.281417 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96968fa5-44ed-4109-8935-684bc644ff19-secret-volume\") pod \"96968fa5-44ed-4109-8935-684bc644ff19\" (UID: \"96968fa5-44ed-4109-8935-684bc644ff19\") " Dec 03 17:30:03 crc kubenswrapper[4998]: I1203 17:30:03.281571 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwbvv\" (UniqueName: \"kubernetes.io/projected/96968fa5-44ed-4109-8935-684bc644ff19-kube-api-access-xwbvv\") pod \"96968fa5-44ed-4109-8935-684bc644ff19\" (UID: \"96968fa5-44ed-4109-8935-684bc644ff19\") " Dec 03 17:30:03 crc kubenswrapper[4998]: I1203 17:30:03.281724 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96968fa5-44ed-4109-8935-684bc644ff19-config-volume\") pod \"96968fa5-44ed-4109-8935-684bc644ff19\" (UID: \"96968fa5-44ed-4109-8935-684bc644ff19\") " Dec 03 17:30:03 crc kubenswrapper[4998]: I1203 17:30:03.283012 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96968fa5-44ed-4109-8935-684bc644ff19-config-volume" (OuterVolumeSpecName: "config-volume") pod "96968fa5-44ed-4109-8935-684bc644ff19" (UID: "96968fa5-44ed-4109-8935-684bc644ff19"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:30:03 crc kubenswrapper[4998]: I1203 17:30:03.288221 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96968fa5-44ed-4109-8935-684bc644ff19-kube-api-access-xwbvv" (OuterVolumeSpecName: "kube-api-access-xwbvv") pod "96968fa5-44ed-4109-8935-684bc644ff19" (UID: "96968fa5-44ed-4109-8935-684bc644ff19"). InnerVolumeSpecName "kube-api-access-xwbvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:30:03 crc kubenswrapper[4998]: I1203 17:30:03.289966 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96968fa5-44ed-4109-8935-684bc644ff19-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "96968fa5-44ed-4109-8935-684bc644ff19" (UID: "96968fa5-44ed-4109-8935-684bc644ff19"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:30:03 crc kubenswrapper[4998]: I1203 17:30:03.384743 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwbvv\" (UniqueName: \"kubernetes.io/projected/96968fa5-44ed-4109-8935-684bc644ff19-kube-api-access-xwbvv\") on node \"crc\" DevicePath \"\"" Dec 03 17:30:03 crc kubenswrapper[4998]: I1203 17:30:03.384840 4998 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96968fa5-44ed-4109-8935-684bc644ff19-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 17:30:03 crc kubenswrapper[4998]: I1203 17:30:03.384862 4998 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96968fa5-44ed-4109-8935-684bc644ff19-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 17:30:03 crc kubenswrapper[4998]: I1203 17:30:03.710384 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-vf5wq" event={"ID":"96968fa5-44ed-4109-8935-684bc644ff19","Type":"ContainerDied","Data":"e01f65ec7243d5e8d4d4bbe242639ed5266f6ec4a258b83ab17357a1be414058"} Dec 03 17:30:03 crc kubenswrapper[4998]: I1203 17:30:03.710468 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e01f65ec7243d5e8d4d4bbe242639ed5266f6ec4a258b83ab17357a1be414058" Dec 03 17:30:03 crc kubenswrapper[4998]: I1203 17:30:03.710473 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-vf5wq" Dec 03 17:30:04 crc kubenswrapper[4998]: I1203 17:30:04.240818 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413005-4z5cj"] Dec 03 17:30:04 crc kubenswrapper[4998]: I1203 17:30:04.252020 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413005-4z5cj"] Dec 03 17:30:05 crc kubenswrapper[4998]: I1203 17:30:05.694238 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51fd9c20-61fa-4b51-913c-6c6745d881e0" path="/var/lib/kubelet/pods/51fd9c20-61fa-4b51-913c-6c6745d881e0/volumes" Dec 03 17:30:06 crc kubenswrapper[4998]: I1203 17:30:06.677455 4998 scope.go:117] "RemoveContainer" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" Dec 03 17:30:06 crc kubenswrapper[4998]: E1203 17:30:06.678072 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:30:18 crc kubenswrapper[4998]: I1203 17:30:18.678207 4998 scope.go:117] "RemoveContainer" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" Dec 03 17:30:18 crc kubenswrapper[4998]: E1203 17:30:18.679212 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:30:30 crc kubenswrapper[4998]: I1203 17:30:30.679043 4998 scope.go:117] "RemoveContainer" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" Dec 03 17:30:31 crc kubenswrapper[4998]: I1203 17:30:31.107440 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"26b784e99bbfd9ccf3b18f20deab0e6b668722269ac44be75e43fa57d4bd8511"} Dec 03 17:31:02 crc kubenswrapper[4998]: I1203 17:31:02.598463 4998 scope.go:117] "RemoveContainer" containerID="6bfc46eb16ba5179dfa693bfce1486f9ef722916c8726a03073ca063f8d44e18" Dec 03 17:32:57 crc kubenswrapper[4998]: I1203 17:32:57.111331 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:32:57 crc kubenswrapper[4998]: I1203 17:32:57.112041 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:33:27 crc kubenswrapper[4998]: I1203 17:33:27.111594 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:33:27 crc kubenswrapper[4998]: I1203 17:33:27.112355 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:33:57 crc kubenswrapper[4998]: I1203 17:33:57.112072 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:33:57 crc kubenswrapper[4998]: I1203 17:33:57.112735 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:33:57 crc kubenswrapper[4998]: I1203 17:33:57.112843 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 17:33:57 crc kubenswrapper[4998]: I1203 17:33:57.113917 4998 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"26b784e99bbfd9ccf3b18f20deab0e6b668722269ac44be75e43fa57d4bd8511"} pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:33:57 crc kubenswrapper[4998]: I1203 17:33:57.114037 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" containerID="cri-o://26b784e99bbfd9ccf3b18f20deab0e6b668722269ac44be75e43fa57d4bd8511" gracePeriod=600 Dec 03 17:33:57 crc kubenswrapper[4998]: I1203 17:33:57.596990 4998 generic.go:334] "Generic (PLEG): container finished" podID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerID="26b784e99bbfd9ccf3b18f20deab0e6b668722269ac44be75e43fa57d4bd8511" exitCode=0 Dec 03 17:33:57 crc kubenswrapper[4998]: I1203 17:33:57.597324 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerDied","Data":"26b784e99bbfd9ccf3b18f20deab0e6b668722269ac44be75e43fa57d4bd8511"} Dec 03 17:33:57 crc kubenswrapper[4998]: I1203 17:33:57.597361 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4"} Dec 03 17:33:57 crc kubenswrapper[4998]: I1203 17:33:57.597386 4998 scope.go:117] "RemoveContainer" containerID="b936dff383566ddb2d088ec8bde9731dddadf89c09c97ed07cbea23378bc4d00" Dec 03 17:35:23 crc kubenswrapper[4998]: I1203 17:35:23.638742 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-htvct"] Dec 03 17:35:23 crc kubenswrapper[4998]: E1203 17:35:23.639862 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96968fa5-44ed-4109-8935-684bc644ff19" containerName="collect-profiles" Dec 03 17:35:23 crc kubenswrapper[4998]: I1203 17:35:23.639878 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="96968fa5-44ed-4109-8935-684bc644ff19" containerName="collect-profiles" Dec 03 17:35:23 crc kubenswrapper[4998]: I1203 17:35:23.640128 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="96968fa5-44ed-4109-8935-684bc644ff19" containerName="collect-profiles" Dec 03 17:35:23 crc kubenswrapper[4998]: I1203 17:35:23.642684 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-htvct" Dec 03 17:35:23 crc kubenswrapper[4998]: I1203 17:35:23.664787 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-htvct"] Dec 03 17:35:23 crc kubenswrapper[4998]: I1203 17:35:23.681097 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4xgk\" (UniqueName: \"kubernetes.io/projected/ca558fea-a4dc-42c1-b83d-4e16afd3d8eb-kube-api-access-n4xgk\") pod \"redhat-operators-htvct\" (UID: \"ca558fea-a4dc-42c1-b83d-4e16afd3d8eb\") " pod="openshift-marketplace/redhat-operators-htvct" Dec 03 17:35:23 crc kubenswrapper[4998]: I1203 17:35:23.681152 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca558fea-a4dc-42c1-b83d-4e16afd3d8eb-catalog-content\") pod \"redhat-operators-htvct\" (UID: \"ca558fea-a4dc-42c1-b83d-4e16afd3d8eb\") " pod="openshift-marketplace/redhat-operators-htvct" Dec 03 17:35:23 crc kubenswrapper[4998]: I1203 17:35:23.681237 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca558fea-a4dc-42c1-b83d-4e16afd3d8eb-utilities\") pod \"redhat-operators-htvct\" (UID: \"ca558fea-a4dc-42c1-b83d-4e16afd3d8eb\") " pod="openshift-marketplace/redhat-operators-htvct" Dec 03 17:35:23 crc kubenswrapper[4998]: I1203 17:35:23.782835 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca558fea-a4dc-42c1-b83d-4e16afd3d8eb-utilities\") pod \"redhat-operators-htvct\" (UID: \"ca558fea-a4dc-42c1-b83d-4e16afd3d8eb\") " pod="openshift-marketplace/redhat-operators-htvct" Dec 03 17:35:23 crc kubenswrapper[4998]: I1203 17:35:23.783089 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4xgk\" (UniqueName: \"kubernetes.io/projected/ca558fea-a4dc-42c1-b83d-4e16afd3d8eb-kube-api-access-n4xgk\") pod \"redhat-operators-htvct\" (UID: \"ca558fea-a4dc-42c1-b83d-4e16afd3d8eb\") " pod="openshift-marketplace/redhat-operators-htvct" Dec 03 17:35:23 crc kubenswrapper[4998]: I1203 17:35:23.783110 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca558fea-a4dc-42c1-b83d-4e16afd3d8eb-catalog-content\") pod \"redhat-operators-htvct\" (UID: \"ca558fea-a4dc-42c1-b83d-4e16afd3d8eb\") " pod="openshift-marketplace/redhat-operators-htvct" Dec 03 17:35:23 crc kubenswrapper[4998]: I1203 17:35:23.784086 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca558fea-a4dc-42c1-b83d-4e16afd3d8eb-catalog-content\") pod \"redhat-operators-htvct\" (UID: \"ca558fea-a4dc-42c1-b83d-4e16afd3d8eb\") " pod="openshift-marketplace/redhat-operators-htvct" Dec 03 17:35:23 crc kubenswrapper[4998]: I1203 17:35:23.784138 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca558fea-a4dc-42c1-b83d-4e16afd3d8eb-utilities\") pod \"redhat-operators-htvct\" (UID: \"ca558fea-a4dc-42c1-b83d-4e16afd3d8eb\") " pod="openshift-marketplace/redhat-operators-htvct" Dec 03 17:35:23 crc kubenswrapper[4998]: I1203 17:35:23.809845 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4xgk\" (UniqueName: \"kubernetes.io/projected/ca558fea-a4dc-42c1-b83d-4e16afd3d8eb-kube-api-access-n4xgk\") pod \"redhat-operators-htvct\" (UID: \"ca558fea-a4dc-42c1-b83d-4e16afd3d8eb\") " pod="openshift-marketplace/redhat-operators-htvct" Dec 03 17:35:23 crc kubenswrapper[4998]: I1203 17:35:23.976983 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-htvct" Dec 03 17:35:24 crc kubenswrapper[4998]: I1203 17:35:24.575018 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-htvct"] Dec 03 17:35:24 crc kubenswrapper[4998]: I1203 17:35:24.642914 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htvct" event={"ID":"ca558fea-a4dc-42c1-b83d-4e16afd3d8eb","Type":"ContainerStarted","Data":"577ead98e7bec1b3c3f8e1c0a8b89274b3bc9a5c5a00684715ea7a545a040df2"} Dec 03 17:35:25 crc kubenswrapper[4998]: I1203 17:35:25.658271 4998 generic.go:334] "Generic (PLEG): container finished" podID="ca558fea-a4dc-42c1-b83d-4e16afd3d8eb" containerID="ebfdd11f4088e9b04bede7901d43048054fa260a26bbcebacbed7ffa0dd1b25f" exitCode=0 Dec 03 17:35:25 crc kubenswrapper[4998]: I1203 17:35:25.658428 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htvct" event={"ID":"ca558fea-a4dc-42c1-b83d-4e16afd3d8eb","Type":"ContainerDied","Data":"ebfdd11f4088e9b04bede7901d43048054fa260a26bbcebacbed7ffa0dd1b25f"} Dec 03 17:35:25 crc kubenswrapper[4998]: I1203 17:35:25.661020 4998 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 17:35:26 crc kubenswrapper[4998]: I1203 17:35:26.671067 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htvct" event={"ID":"ca558fea-a4dc-42c1-b83d-4e16afd3d8eb","Type":"ContainerStarted","Data":"c9c85fcf851bc29fc2f9fb30304caa56022524b3943dc2be71b048e740e96d92"} Dec 03 17:35:30 crc kubenswrapper[4998]: I1203 17:35:30.710107 4998 generic.go:334] "Generic (PLEG): container finished" podID="ca558fea-a4dc-42c1-b83d-4e16afd3d8eb" containerID="c9c85fcf851bc29fc2f9fb30304caa56022524b3943dc2be71b048e740e96d92" exitCode=0 Dec 03 17:35:30 crc kubenswrapper[4998]: I1203 17:35:30.710185 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htvct" event={"ID":"ca558fea-a4dc-42c1-b83d-4e16afd3d8eb","Type":"ContainerDied","Data":"c9c85fcf851bc29fc2f9fb30304caa56022524b3943dc2be71b048e740e96d92"} Dec 03 17:35:32 crc kubenswrapper[4998]: I1203 17:35:32.733175 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htvct" event={"ID":"ca558fea-a4dc-42c1-b83d-4e16afd3d8eb","Type":"ContainerStarted","Data":"2ddc951b7a87bccf82e70a6c6d5e3a2e9ff92f20d5717fe686efddf0c1d8229d"} Dec 03 17:35:32 crc kubenswrapper[4998]: I1203 17:35:32.782042 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-htvct" podStartSLOduration=4.112349111 podStartE2EDuration="9.782011412s" podCreationTimestamp="2025-12-03 17:35:23 +0000 UTC" firstStartedPulling="2025-12-03 17:35:25.660832277 +0000 UTC m=+5504.272532500" lastFinishedPulling="2025-12-03 17:35:31.330494578 +0000 UTC m=+5509.942194801" observedRunningTime="2025-12-03 17:35:32.761884942 +0000 UTC m=+5511.373585165" watchObservedRunningTime="2025-12-03 17:35:32.782011412 +0000 UTC m=+5511.393711655" Dec 03 17:35:33 crc kubenswrapper[4998]: I1203 17:35:33.977388 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-htvct" Dec 03 17:35:33 crc kubenswrapper[4998]: I1203 17:35:33.977865 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-htvct" Dec 03 17:35:35 crc kubenswrapper[4998]: I1203 17:35:35.051452 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-htvct" podUID="ca558fea-a4dc-42c1-b83d-4e16afd3d8eb" containerName="registry-server" probeResult="failure" output=< Dec 03 17:35:35 crc kubenswrapper[4998]: timeout: failed to connect service ":50051" within 1s Dec 03 17:35:35 crc kubenswrapper[4998]: > Dec 03 17:35:44 crc kubenswrapper[4998]: I1203 17:35:44.026562 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-htvct" Dec 03 17:35:44 crc kubenswrapper[4998]: I1203 17:35:44.071650 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-htvct" Dec 03 17:35:44 crc kubenswrapper[4998]: I1203 17:35:44.262622 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-htvct"] Dec 03 17:35:45 crc kubenswrapper[4998]: I1203 17:35:45.872803 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-htvct" podUID="ca558fea-a4dc-42c1-b83d-4e16afd3d8eb" containerName="registry-server" containerID="cri-o://2ddc951b7a87bccf82e70a6c6d5e3a2e9ff92f20d5717fe686efddf0c1d8229d" gracePeriod=2 Dec 03 17:35:46 crc kubenswrapper[4998]: I1203 17:35:46.894478 4998 generic.go:334] "Generic (PLEG): container finished" podID="ca558fea-a4dc-42c1-b83d-4e16afd3d8eb" containerID="2ddc951b7a87bccf82e70a6c6d5e3a2e9ff92f20d5717fe686efddf0c1d8229d" exitCode=0 Dec 03 17:35:46 crc kubenswrapper[4998]: I1203 17:35:46.894713 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htvct" event={"ID":"ca558fea-a4dc-42c1-b83d-4e16afd3d8eb","Type":"ContainerDied","Data":"2ddc951b7a87bccf82e70a6c6d5e3a2e9ff92f20d5717fe686efddf0c1d8229d"} Dec 03 17:35:46 crc kubenswrapper[4998]: I1203 17:35:46.894888 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-htvct" event={"ID":"ca558fea-a4dc-42c1-b83d-4e16afd3d8eb","Type":"ContainerDied","Data":"577ead98e7bec1b3c3f8e1c0a8b89274b3bc9a5c5a00684715ea7a545a040df2"} Dec 03 17:35:46 crc kubenswrapper[4998]: I1203 17:35:46.894903 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="577ead98e7bec1b3c3f8e1c0a8b89274b3bc9a5c5a00684715ea7a545a040df2" Dec 03 17:35:46 crc kubenswrapper[4998]: I1203 17:35:46.926139 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-htvct" Dec 03 17:35:47 crc kubenswrapper[4998]: I1203 17:35:47.035646 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4xgk\" (UniqueName: \"kubernetes.io/projected/ca558fea-a4dc-42c1-b83d-4e16afd3d8eb-kube-api-access-n4xgk\") pod \"ca558fea-a4dc-42c1-b83d-4e16afd3d8eb\" (UID: \"ca558fea-a4dc-42c1-b83d-4e16afd3d8eb\") " Dec 03 17:35:47 crc kubenswrapper[4998]: I1203 17:35:47.036043 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca558fea-a4dc-42c1-b83d-4e16afd3d8eb-utilities\") pod \"ca558fea-a4dc-42c1-b83d-4e16afd3d8eb\" (UID: \"ca558fea-a4dc-42c1-b83d-4e16afd3d8eb\") " Dec 03 17:35:47 crc kubenswrapper[4998]: I1203 17:35:47.036091 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca558fea-a4dc-42c1-b83d-4e16afd3d8eb-catalog-content\") pod \"ca558fea-a4dc-42c1-b83d-4e16afd3d8eb\" (UID: \"ca558fea-a4dc-42c1-b83d-4e16afd3d8eb\") " Dec 03 17:35:47 crc kubenswrapper[4998]: I1203 17:35:47.036919 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca558fea-a4dc-42c1-b83d-4e16afd3d8eb-utilities" (OuterVolumeSpecName: "utilities") pod "ca558fea-a4dc-42c1-b83d-4e16afd3d8eb" (UID: "ca558fea-a4dc-42c1-b83d-4e16afd3d8eb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:35:47 crc kubenswrapper[4998]: I1203 17:35:47.042923 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca558fea-a4dc-42c1-b83d-4e16afd3d8eb-kube-api-access-n4xgk" (OuterVolumeSpecName: "kube-api-access-n4xgk") pod "ca558fea-a4dc-42c1-b83d-4e16afd3d8eb" (UID: "ca558fea-a4dc-42c1-b83d-4e16afd3d8eb"). InnerVolumeSpecName "kube-api-access-n4xgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:47 crc kubenswrapper[4998]: I1203 17:35:47.138869 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca558fea-a4dc-42c1-b83d-4e16afd3d8eb-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:47 crc kubenswrapper[4998]: I1203 17:35:47.138914 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4xgk\" (UniqueName: \"kubernetes.io/projected/ca558fea-a4dc-42c1-b83d-4e16afd3d8eb-kube-api-access-n4xgk\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:47 crc kubenswrapper[4998]: I1203 17:35:47.187906 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca558fea-a4dc-42c1-b83d-4e16afd3d8eb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ca558fea-a4dc-42c1-b83d-4e16afd3d8eb" (UID: "ca558fea-a4dc-42c1-b83d-4e16afd3d8eb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:35:47 crc kubenswrapper[4998]: I1203 17:35:47.241702 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca558fea-a4dc-42c1-b83d-4e16afd3d8eb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:47 crc kubenswrapper[4998]: I1203 17:35:47.904476 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-htvct" Dec 03 17:35:47 crc kubenswrapper[4998]: I1203 17:35:47.946539 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-htvct"] Dec 03 17:35:47 crc kubenswrapper[4998]: I1203 17:35:47.962725 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-htvct"] Dec 03 17:35:49 crc kubenswrapper[4998]: I1203 17:35:49.692254 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca558fea-a4dc-42c1-b83d-4e16afd3d8eb" path="/var/lib/kubelet/pods/ca558fea-a4dc-42c1-b83d-4e16afd3d8eb/volumes" Dec 03 17:35:57 crc kubenswrapper[4998]: I1203 17:35:57.110890 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:35:57 crc kubenswrapper[4998]: I1203 17:35:57.111398 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:36:27 crc kubenswrapper[4998]: I1203 17:36:27.111168 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:36:27 crc kubenswrapper[4998]: I1203 17:36:27.111741 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:36:57 crc kubenswrapper[4998]: I1203 17:36:57.110698 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:36:57 crc kubenswrapper[4998]: I1203 17:36:57.111148 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:36:57 crc kubenswrapper[4998]: I1203 17:36:57.111191 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 17:36:57 crc kubenswrapper[4998]: I1203 17:36:57.111957 4998 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4"} pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:36:57 crc kubenswrapper[4998]: I1203 17:36:57.112010 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" containerID="cri-o://7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" gracePeriod=600 Dec 03 17:36:57 crc kubenswrapper[4998]: E1203 17:36:57.245255 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:36:57 crc kubenswrapper[4998]: I1203 17:36:57.699307 4998 generic.go:334] "Generic (PLEG): container finished" podID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" exitCode=0 Dec 03 17:36:57 crc kubenswrapper[4998]: I1203 17:36:57.699361 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerDied","Data":"7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4"} Dec 03 17:36:57 crc kubenswrapper[4998]: I1203 17:36:57.699407 4998 scope.go:117] "RemoveContainer" containerID="26b784e99bbfd9ccf3b18f20deab0e6b668722269ac44be75e43fa57d4bd8511" Dec 03 17:36:57 crc kubenswrapper[4998]: I1203 17:36:57.700170 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:36:57 crc kubenswrapper[4998]: E1203 17:36:57.700612 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:37:10 crc kubenswrapper[4998]: I1203 17:37:10.678237 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:37:10 crc kubenswrapper[4998]: E1203 17:37:10.679153 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:37:23 crc kubenswrapper[4998]: I1203 17:37:23.682154 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:37:23 crc kubenswrapper[4998]: E1203 17:37:23.683733 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:37:38 crc kubenswrapper[4998]: I1203 17:37:38.677978 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:37:38 crc kubenswrapper[4998]: E1203 17:37:38.678992 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:37:53 crc kubenswrapper[4998]: I1203 17:37:53.685362 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:37:53 crc kubenswrapper[4998]: E1203 17:37:53.686337 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:38:04 crc kubenswrapper[4998]: I1203 17:38:04.677925 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:38:04 crc kubenswrapper[4998]: E1203 17:38:04.678805 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:38:18 crc kubenswrapper[4998]: I1203 17:38:18.678155 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:38:18 crc kubenswrapper[4998]: E1203 17:38:18.680524 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:38:29 crc kubenswrapper[4998]: I1203 17:38:29.680395 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:38:29 crc kubenswrapper[4998]: E1203 17:38:29.682120 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:38:43 crc kubenswrapper[4998]: I1203 17:38:43.694566 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:38:43 crc kubenswrapper[4998]: E1203 17:38:43.695588 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:38:53 crc kubenswrapper[4998]: I1203 17:38:53.785910 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zwrmj"] Dec 03 17:38:53 crc kubenswrapper[4998]: E1203 17:38:53.787373 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca558fea-a4dc-42c1-b83d-4e16afd3d8eb" containerName="registry-server" Dec 03 17:38:53 crc kubenswrapper[4998]: I1203 17:38:53.787401 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca558fea-a4dc-42c1-b83d-4e16afd3d8eb" containerName="registry-server" Dec 03 17:38:53 crc kubenswrapper[4998]: E1203 17:38:53.787491 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca558fea-a4dc-42c1-b83d-4e16afd3d8eb" containerName="extract-content" Dec 03 17:38:53 crc kubenswrapper[4998]: I1203 17:38:53.787509 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca558fea-a4dc-42c1-b83d-4e16afd3d8eb" containerName="extract-content" Dec 03 17:38:53 crc kubenswrapper[4998]: E1203 17:38:53.787550 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca558fea-a4dc-42c1-b83d-4e16afd3d8eb" containerName="extract-utilities" Dec 03 17:38:53 crc kubenswrapper[4998]: I1203 17:38:53.787563 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca558fea-a4dc-42c1-b83d-4e16afd3d8eb" containerName="extract-utilities" Dec 03 17:38:53 crc kubenswrapper[4998]: I1203 17:38:53.787990 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca558fea-a4dc-42c1-b83d-4e16afd3d8eb" containerName="registry-server" Dec 03 17:38:53 crc kubenswrapper[4998]: I1203 17:38:53.791712 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zwrmj" Dec 03 17:38:53 crc kubenswrapper[4998]: I1203 17:38:53.794256 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zwrmj"] Dec 03 17:38:53 crc kubenswrapper[4998]: I1203 17:38:53.882812 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/003fbeb2-e822-416e-991a-62abdc7afe87-utilities\") pod \"redhat-marketplace-zwrmj\" (UID: \"003fbeb2-e822-416e-991a-62abdc7afe87\") " pod="openshift-marketplace/redhat-marketplace-zwrmj" Dec 03 17:38:53 crc kubenswrapper[4998]: I1203 17:38:53.883325 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/003fbeb2-e822-416e-991a-62abdc7afe87-catalog-content\") pod \"redhat-marketplace-zwrmj\" (UID: \"003fbeb2-e822-416e-991a-62abdc7afe87\") " pod="openshift-marketplace/redhat-marketplace-zwrmj" Dec 03 17:38:53 crc kubenswrapper[4998]: I1203 17:38:53.883356 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chq92\" (UniqueName: \"kubernetes.io/projected/003fbeb2-e822-416e-991a-62abdc7afe87-kube-api-access-chq92\") pod \"redhat-marketplace-zwrmj\" (UID: \"003fbeb2-e822-416e-991a-62abdc7afe87\") " pod="openshift-marketplace/redhat-marketplace-zwrmj" Dec 03 17:38:53 crc kubenswrapper[4998]: I1203 17:38:53.985595 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/003fbeb2-e822-416e-991a-62abdc7afe87-catalog-content\") pod \"redhat-marketplace-zwrmj\" (UID: \"003fbeb2-e822-416e-991a-62abdc7afe87\") " pod="openshift-marketplace/redhat-marketplace-zwrmj" Dec 03 17:38:53 crc kubenswrapper[4998]: I1203 17:38:53.985650 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chq92\" (UniqueName: \"kubernetes.io/projected/003fbeb2-e822-416e-991a-62abdc7afe87-kube-api-access-chq92\") pod \"redhat-marketplace-zwrmj\" (UID: \"003fbeb2-e822-416e-991a-62abdc7afe87\") " pod="openshift-marketplace/redhat-marketplace-zwrmj" Dec 03 17:38:53 crc kubenswrapper[4998]: I1203 17:38:53.985790 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/003fbeb2-e822-416e-991a-62abdc7afe87-utilities\") pod \"redhat-marketplace-zwrmj\" (UID: \"003fbeb2-e822-416e-991a-62abdc7afe87\") " pod="openshift-marketplace/redhat-marketplace-zwrmj" Dec 03 17:38:53 crc kubenswrapper[4998]: I1203 17:38:53.986155 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/003fbeb2-e822-416e-991a-62abdc7afe87-catalog-content\") pod \"redhat-marketplace-zwrmj\" (UID: \"003fbeb2-e822-416e-991a-62abdc7afe87\") " pod="openshift-marketplace/redhat-marketplace-zwrmj" Dec 03 17:38:53 crc kubenswrapper[4998]: I1203 17:38:53.986788 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/003fbeb2-e822-416e-991a-62abdc7afe87-utilities\") pod \"redhat-marketplace-zwrmj\" (UID: \"003fbeb2-e822-416e-991a-62abdc7afe87\") " pod="openshift-marketplace/redhat-marketplace-zwrmj" Dec 03 17:38:54 crc kubenswrapper[4998]: I1203 17:38:54.010877 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chq92\" (UniqueName: \"kubernetes.io/projected/003fbeb2-e822-416e-991a-62abdc7afe87-kube-api-access-chq92\") pod \"redhat-marketplace-zwrmj\" (UID: \"003fbeb2-e822-416e-991a-62abdc7afe87\") " pod="openshift-marketplace/redhat-marketplace-zwrmj" Dec 03 17:38:54 crc kubenswrapper[4998]: I1203 17:38:54.228674 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zwrmj" Dec 03 17:38:54 crc kubenswrapper[4998]: I1203 17:38:54.976970 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zwrmj"] Dec 03 17:38:54 crc kubenswrapper[4998]: W1203 17:38:54.982136 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod003fbeb2_e822_416e_991a_62abdc7afe87.slice/crio-3a581f6c32173cc017370002b48c542e689469980760e35a2b75a6b60596cd7b WatchSource:0}: Error finding container 3a581f6c32173cc017370002b48c542e689469980760e35a2b75a6b60596cd7b: Status 404 returned error can't find the container with id 3a581f6c32173cc017370002b48c542e689469980760e35a2b75a6b60596cd7b Dec 03 17:38:55 crc kubenswrapper[4998]: I1203 17:38:55.263273 4998 generic.go:334] "Generic (PLEG): container finished" podID="003fbeb2-e822-416e-991a-62abdc7afe87" containerID="8632e9f33101486660e76bcdc29c143ad5320af9eab5eda6f56431d4cd5a929b" exitCode=0 Dec 03 17:38:55 crc kubenswrapper[4998]: I1203 17:38:55.263379 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zwrmj" event={"ID":"003fbeb2-e822-416e-991a-62abdc7afe87","Type":"ContainerDied","Data":"8632e9f33101486660e76bcdc29c143ad5320af9eab5eda6f56431d4cd5a929b"} Dec 03 17:38:55 crc kubenswrapper[4998]: I1203 17:38:55.263617 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zwrmj" event={"ID":"003fbeb2-e822-416e-991a-62abdc7afe87","Type":"ContainerStarted","Data":"3a581f6c32173cc017370002b48c542e689469980760e35a2b75a6b60596cd7b"} Dec 03 17:38:56 crc kubenswrapper[4998]: I1203 17:38:56.278337 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zwrmj" event={"ID":"003fbeb2-e822-416e-991a-62abdc7afe87","Type":"ContainerStarted","Data":"d2005b7a05103d7c087e3e98cf4c9421a9f958a5e5b30edbd1b22602d136d581"} Dec 03 17:38:57 crc kubenswrapper[4998]: I1203 17:38:57.295533 4998 generic.go:334] "Generic (PLEG): container finished" podID="003fbeb2-e822-416e-991a-62abdc7afe87" containerID="d2005b7a05103d7c087e3e98cf4c9421a9f958a5e5b30edbd1b22602d136d581" exitCode=0 Dec 03 17:38:57 crc kubenswrapper[4998]: I1203 17:38:57.295580 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zwrmj" event={"ID":"003fbeb2-e822-416e-991a-62abdc7afe87","Type":"ContainerDied","Data":"d2005b7a05103d7c087e3e98cf4c9421a9f958a5e5b30edbd1b22602d136d581"} Dec 03 17:38:58 crc kubenswrapper[4998]: I1203 17:38:58.310741 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zwrmj" event={"ID":"003fbeb2-e822-416e-991a-62abdc7afe87","Type":"ContainerStarted","Data":"6399760b8212a810eae535b7e6c8380d00d0924caba81dc2a9a88bc59bc9d5f7"} Dec 03 17:38:58 crc kubenswrapper[4998]: I1203 17:38:58.336806 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zwrmj" podStartSLOduration=2.646633962 podStartE2EDuration="5.336787874s" podCreationTimestamp="2025-12-03 17:38:53 +0000 UTC" firstStartedPulling="2025-12-03 17:38:55.265316355 +0000 UTC m=+5713.877016578" lastFinishedPulling="2025-12-03 17:38:57.955470257 +0000 UTC m=+5716.567170490" observedRunningTime="2025-12-03 17:38:58.333137003 +0000 UTC m=+5716.944837236" watchObservedRunningTime="2025-12-03 17:38:58.336787874 +0000 UTC m=+5716.948488117" Dec 03 17:38:58 crc kubenswrapper[4998]: I1203 17:38:58.678138 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:38:58 crc kubenswrapper[4998]: E1203 17:38:58.678823 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:39:04 crc kubenswrapper[4998]: I1203 17:39:04.229839 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zwrmj" Dec 03 17:39:04 crc kubenswrapper[4998]: I1203 17:39:04.230685 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zwrmj" Dec 03 17:39:04 crc kubenswrapper[4998]: I1203 17:39:04.297252 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zwrmj" Dec 03 17:39:04 crc kubenswrapper[4998]: I1203 17:39:04.434867 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zwrmj" Dec 03 17:39:04 crc kubenswrapper[4998]: I1203 17:39:04.537353 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zwrmj"] Dec 03 17:39:06 crc kubenswrapper[4998]: I1203 17:39:06.398100 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zwrmj" podUID="003fbeb2-e822-416e-991a-62abdc7afe87" containerName="registry-server" containerID="cri-o://6399760b8212a810eae535b7e6c8380d00d0924caba81dc2a9a88bc59bc9d5f7" gracePeriod=2 Dec 03 17:39:06 crc kubenswrapper[4998]: E1203 17:39:06.746539 4998 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod003fbeb2_e822_416e_991a_62abdc7afe87.slice/crio-6399760b8212a810eae535b7e6c8380d00d0924caba81dc2a9a88bc59bc9d5f7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod003fbeb2_e822_416e_991a_62abdc7afe87.slice/crio-conmon-6399760b8212a810eae535b7e6c8380d00d0924caba81dc2a9a88bc59bc9d5f7.scope\": RecentStats: unable to find data in memory cache]" Dec 03 17:39:06 crc kubenswrapper[4998]: I1203 17:39:06.992964 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zwrmj" Dec 03 17:39:07 crc kubenswrapper[4998]: I1203 17:39:07.023869 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/003fbeb2-e822-416e-991a-62abdc7afe87-catalog-content\") pod \"003fbeb2-e822-416e-991a-62abdc7afe87\" (UID: \"003fbeb2-e822-416e-991a-62abdc7afe87\") " Dec 03 17:39:07 crc kubenswrapper[4998]: I1203 17:39:07.023922 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/003fbeb2-e822-416e-991a-62abdc7afe87-utilities\") pod \"003fbeb2-e822-416e-991a-62abdc7afe87\" (UID: \"003fbeb2-e822-416e-991a-62abdc7afe87\") " Dec 03 17:39:07 crc kubenswrapper[4998]: I1203 17:39:07.024051 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chq92\" (UniqueName: \"kubernetes.io/projected/003fbeb2-e822-416e-991a-62abdc7afe87-kube-api-access-chq92\") pod \"003fbeb2-e822-416e-991a-62abdc7afe87\" (UID: \"003fbeb2-e822-416e-991a-62abdc7afe87\") " Dec 03 17:39:07 crc kubenswrapper[4998]: I1203 17:39:07.025959 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/003fbeb2-e822-416e-991a-62abdc7afe87-utilities" (OuterVolumeSpecName: "utilities") pod "003fbeb2-e822-416e-991a-62abdc7afe87" (UID: "003fbeb2-e822-416e-991a-62abdc7afe87"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:39:07 crc kubenswrapper[4998]: I1203 17:39:07.034359 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/003fbeb2-e822-416e-991a-62abdc7afe87-kube-api-access-chq92" (OuterVolumeSpecName: "kube-api-access-chq92") pod "003fbeb2-e822-416e-991a-62abdc7afe87" (UID: "003fbeb2-e822-416e-991a-62abdc7afe87"). InnerVolumeSpecName "kube-api-access-chq92". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:39:07 crc kubenswrapper[4998]: I1203 17:39:07.060158 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/003fbeb2-e822-416e-991a-62abdc7afe87-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "003fbeb2-e822-416e-991a-62abdc7afe87" (UID: "003fbeb2-e822-416e-991a-62abdc7afe87"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:39:07 crc kubenswrapper[4998]: I1203 17:39:07.126146 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chq92\" (UniqueName: \"kubernetes.io/projected/003fbeb2-e822-416e-991a-62abdc7afe87-kube-api-access-chq92\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:07 crc kubenswrapper[4998]: I1203 17:39:07.126199 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/003fbeb2-e822-416e-991a-62abdc7afe87-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:07 crc kubenswrapper[4998]: I1203 17:39:07.126219 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/003fbeb2-e822-416e-991a-62abdc7afe87-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:07 crc kubenswrapper[4998]: I1203 17:39:07.412709 4998 generic.go:334] "Generic (PLEG): container finished" podID="003fbeb2-e822-416e-991a-62abdc7afe87" containerID="6399760b8212a810eae535b7e6c8380d00d0924caba81dc2a9a88bc59bc9d5f7" exitCode=0 Dec 03 17:39:07 crc kubenswrapper[4998]: I1203 17:39:07.412778 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zwrmj" event={"ID":"003fbeb2-e822-416e-991a-62abdc7afe87","Type":"ContainerDied","Data":"6399760b8212a810eae535b7e6c8380d00d0924caba81dc2a9a88bc59bc9d5f7"} Dec 03 17:39:07 crc kubenswrapper[4998]: I1203 17:39:07.412822 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zwrmj" event={"ID":"003fbeb2-e822-416e-991a-62abdc7afe87","Type":"ContainerDied","Data":"3a581f6c32173cc017370002b48c542e689469980760e35a2b75a6b60596cd7b"} Dec 03 17:39:07 crc kubenswrapper[4998]: I1203 17:39:07.412840 4998 scope.go:117] "RemoveContainer" containerID="6399760b8212a810eae535b7e6c8380d00d0924caba81dc2a9a88bc59bc9d5f7" Dec 03 17:39:07 crc kubenswrapper[4998]: I1203 17:39:07.412846 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zwrmj" Dec 03 17:39:07 crc kubenswrapper[4998]: I1203 17:39:07.457402 4998 scope.go:117] "RemoveContainer" containerID="d2005b7a05103d7c087e3e98cf4c9421a9f958a5e5b30edbd1b22602d136d581" Dec 03 17:39:07 crc kubenswrapper[4998]: I1203 17:39:07.476354 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zwrmj"] Dec 03 17:39:07 crc kubenswrapper[4998]: I1203 17:39:07.509527 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zwrmj"] Dec 03 17:39:07 crc kubenswrapper[4998]: I1203 17:39:07.526253 4998 scope.go:117] "RemoveContainer" containerID="8632e9f33101486660e76bcdc29c143ad5320af9eab5eda6f56431d4cd5a929b" Dec 03 17:39:07 crc kubenswrapper[4998]: I1203 17:39:07.582525 4998 scope.go:117] "RemoveContainer" containerID="6399760b8212a810eae535b7e6c8380d00d0924caba81dc2a9a88bc59bc9d5f7" Dec 03 17:39:07 crc kubenswrapper[4998]: E1203 17:39:07.583135 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6399760b8212a810eae535b7e6c8380d00d0924caba81dc2a9a88bc59bc9d5f7\": container with ID starting with 6399760b8212a810eae535b7e6c8380d00d0924caba81dc2a9a88bc59bc9d5f7 not found: ID does not exist" containerID="6399760b8212a810eae535b7e6c8380d00d0924caba81dc2a9a88bc59bc9d5f7" Dec 03 17:39:07 crc kubenswrapper[4998]: I1203 17:39:07.583186 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6399760b8212a810eae535b7e6c8380d00d0924caba81dc2a9a88bc59bc9d5f7"} err="failed to get container status \"6399760b8212a810eae535b7e6c8380d00d0924caba81dc2a9a88bc59bc9d5f7\": rpc error: code = NotFound desc = could not find container \"6399760b8212a810eae535b7e6c8380d00d0924caba81dc2a9a88bc59bc9d5f7\": container with ID starting with 6399760b8212a810eae535b7e6c8380d00d0924caba81dc2a9a88bc59bc9d5f7 not found: ID does not exist" Dec 03 17:39:07 crc kubenswrapper[4998]: I1203 17:39:07.583213 4998 scope.go:117] "RemoveContainer" containerID="d2005b7a05103d7c087e3e98cf4c9421a9f958a5e5b30edbd1b22602d136d581" Dec 03 17:39:07 crc kubenswrapper[4998]: E1203 17:39:07.583654 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2005b7a05103d7c087e3e98cf4c9421a9f958a5e5b30edbd1b22602d136d581\": container with ID starting with d2005b7a05103d7c087e3e98cf4c9421a9f958a5e5b30edbd1b22602d136d581 not found: ID does not exist" containerID="d2005b7a05103d7c087e3e98cf4c9421a9f958a5e5b30edbd1b22602d136d581" Dec 03 17:39:07 crc kubenswrapper[4998]: I1203 17:39:07.583808 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2005b7a05103d7c087e3e98cf4c9421a9f958a5e5b30edbd1b22602d136d581"} err="failed to get container status \"d2005b7a05103d7c087e3e98cf4c9421a9f958a5e5b30edbd1b22602d136d581\": rpc error: code = NotFound desc = could not find container \"d2005b7a05103d7c087e3e98cf4c9421a9f958a5e5b30edbd1b22602d136d581\": container with ID starting with d2005b7a05103d7c087e3e98cf4c9421a9f958a5e5b30edbd1b22602d136d581 not found: ID does not exist" Dec 03 17:39:07 crc kubenswrapper[4998]: I1203 17:39:07.583914 4998 scope.go:117] "RemoveContainer" containerID="8632e9f33101486660e76bcdc29c143ad5320af9eab5eda6f56431d4cd5a929b" Dec 03 17:39:07 crc kubenswrapper[4998]: E1203 17:39:07.584314 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8632e9f33101486660e76bcdc29c143ad5320af9eab5eda6f56431d4cd5a929b\": container with ID starting with 8632e9f33101486660e76bcdc29c143ad5320af9eab5eda6f56431d4cd5a929b not found: ID does not exist" containerID="8632e9f33101486660e76bcdc29c143ad5320af9eab5eda6f56431d4cd5a929b" Dec 03 17:39:07 crc kubenswrapper[4998]: I1203 17:39:07.584339 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8632e9f33101486660e76bcdc29c143ad5320af9eab5eda6f56431d4cd5a929b"} err="failed to get container status \"8632e9f33101486660e76bcdc29c143ad5320af9eab5eda6f56431d4cd5a929b\": rpc error: code = NotFound desc = could not find container \"8632e9f33101486660e76bcdc29c143ad5320af9eab5eda6f56431d4cd5a929b\": container with ID starting with 8632e9f33101486660e76bcdc29c143ad5320af9eab5eda6f56431d4cd5a929b not found: ID does not exist" Dec 03 17:39:07 crc kubenswrapper[4998]: I1203 17:39:07.692946 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="003fbeb2-e822-416e-991a-62abdc7afe87" path="/var/lib/kubelet/pods/003fbeb2-e822-416e-991a-62abdc7afe87/volumes" Dec 03 17:39:10 crc kubenswrapper[4998]: I1203 17:39:10.679907 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:39:10 crc kubenswrapper[4998]: E1203 17:39:10.680928 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:39:21 crc kubenswrapper[4998]: I1203 17:39:21.684391 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:39:21 crc kubenswrapper[4998]: E1203 17:39:21.685136 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:39:32 crc kubenswrapper[4998]: I1203 17:39:32.678169 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:39:32 crc kubenswrapper[4998]: E1203 17:39:32.679158 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:39:45 crc kubenswrapper[4998]: I1203 17:39:45.679931 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:39:45 crc kubenswrapper[4998]: E1203 17:39:45.681419 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:40:00 crc kubenswrapper[4998]: I1203 17:40:00.678117 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:40:00 crc kubenswrapper[4998]: E1203 17:40:00.681452 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:40:07 crc kubenswrapper[4998]: I1203 17:40:07.856458 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mqh5v"] Dec 03 17:40:07 crc kubenswrapper[4998]: E1203 17:40:07.857338 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="003fbeb2-e822-416e-991a-62abdc7afe87" containerName="extract-content" Dec 03 17:40:07 crc kubenswrapper[4998]: I1203 17:40:07.857350 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="003fbeb2-e822-416e-991a-62abdc7afe87" containerName="extract-content" Dec 03 17:40:07 crc kubenswrapper[4998]: E1203 17:40:07.857365 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="003fbeb2-e822-416e-991a-62abdc7afe87" containerName="extract-utilities" Dec 03 17:40:07 crc kubenswrapper[4998]: I1203 17:40:07.857371 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="003fbeb2-e822-416e-991a-62abdc7afe87" containerName="extract-utilities" Dec 03 17:40:07 crc kubenswrapper[4998]: E1203 17:40:07.857399 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="003fbeb2-e822-416e-991a-62abdc7afe87" containerName="registry-server" Dec 03 17:40:07 crc kubenswrapper[4998]: I1203 17:40:07.857408 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="003fbeb2-e822-416e-991a-62abdc7afe87" containerName="registry-server" Dec 03 17:40:07 crc kubenswrapper[4998]: I1203 17:40:07.857609 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="003fbeb2-e822-416e-991a-62abdc7afe87" containerName="registry-server" Dec 03 17:40:07 crc kubenswrapper[4998]: I1203 17:40:07.860795 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mqh5v" Dec 03 17:40:07 crc kubenswrapper[4998]: I1203 17:40:07.878929 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4eed4b9a-97db-42f1-b43b-1b86dcb28b80-catalog-content\") pod \"community-operators-mqh5v\" (UID: \"4eed4b9a-97db-42f1-b43b-1b86dcb28b80\") " pod="openshift-marketplace/community-operators-mqh5v" Dec 03 17:40:07 crc kubenswrapper[4998]: I1203 17:40:07.879055 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4eed4b9a-97db-42f1-b43b-1b86dcb28b80-utilities\") pod \"community-operators-mqh5v\" (UID: \"4eed4b9a-97db-42f1-b43b-1b86dcb28b80\") " pod="openshift-marketplace/community-operators-mqh5v" Dec 03 17:40:07 crc kubenswrapper[4998]: I1203 17:40:07.879125 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdvwd\" (UniqueName: \"kubernetes.io/projected/4eed4b9a-97db-42f1-b43b-1b86dcb28b80-kube-api-access-wdvwd\") pod \"community-operators-mqh5v\" (UID: \"4eed4b9a-97db-42f1-b43b-1b86dcb28b80\") " pod="openshift-marketplace/community-operators-mqh5v" Dec 03 17:40:07 crc kubenswrapper[4998]: I1203 17:40:07.887994 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mqh5v"] Dec 03 17:40:07 crc kubenswrapper[4998]: I1203 17:40:07.980191 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4eed4b9a-97db-42f1-b43b-1b86dcb28b80-utilities\") pod \"community-operators-mqh5v\" (UID: \"4eed4b9a-97db-42f1-b43b-1b86dcb28b80\") " pod="openshift-marketplace/community-operators-mqh5v" Dec 03 17:40:07 crc kubenswrapper[4998]: I1203 17:40:07.980284 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdvwd\" (UniqueName: \"kubernetes.io/projected/4eed4b9a-97db-42f1-b43b-1b86dcb28b80-kube-api-access-wdvwd\") pod \"community-operators-mqh5v\" (UID: \"4eed4b9a-97db-42f1-b43b-1b86dcb28b80\") " pod="openshift-marketplace/community-operators-mqh5v" Dec 03 17:40:07 crc kubenswrapper[4998]: I1203 17:40:07.980371 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4eed4b9a-97db-42f1-b43b-1b86dcb28b80-catalog-content\") pod \"community-operators-mqh5v\" (UID: \"4eed4b9a-97db-42f1-b43b-1b86dcb28b80\") " pod="openshift-marketplace/community-operators-mqh5v" Dec 03 17:40:07 crc kubenswrapper[4998]: I1203 17:40:07.980839 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4eed4b9a-97db-42f1-b43b-1b86dcb28b80-utilities\") pod \"community-operators-mqh5v\" (UID: \"4eed4b9a-97db-42f1-b43b-1b86dcb28b80\") " pod="openshift-marketplace/community-operators-mqh5v" Dec 03 17:40:07 crc kubenswrapper[4998]: I1203 17:40:07.980859 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4eed4b9a-97db-42f1-b43b-1b86dcb28b80-catalog-content\") pod \"community-operators-mqh5v\" (UID: \"4eed4b9a-97db-42f1-b43b-1b86dcb28b80\") " pod="openshift-marketplace/community-operators-mqh5v" Dec 03 17:40:08 crc kubenswrapper[4998]: I1203 17:40:08.005683 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdvwd\" (UniqueName: \"kubernetes.io/projected/4eed4b9a-97db-42f1-b43b-1b86dcb28b80-kube-api-access-wdvwd\") pod \"community-operators-mqh5v\" (UID: \"4eed4b9a-97db-42f1-b43b-1b86dcb28b80\") " pod="openshift-marketplace/community-operators-mqh5v" Dec 03 17:40:08 crc kubenswrapper[4998]: I1203 17:40:08.178126 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mqh5v" Dec 03 17:40:08 crc kubenswrapper[4998]: I1203 17:40:08.688612 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mqh5v"] Dec 03 17:40:09 crc kubenswrapper[4998]: I1203 17:40:09.156716 4998 generic.go:334] "Generic (PLEG): container finished" podID="4eed4b9a-97db-42f1-b43b-1b86dcb28b80" containerID="7c7d717b1e29083a433e71613e0683a5404c9b376452940760f7722e834726ff" exitCode=0 Dec 03 17:40:09 crc kubenswrapper[4998]: I1203 17:40:09.156802 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mqh5v" event={"ID":"4eed4b9a-97db-42f1-b43b-1b86dcb28b80","Type":"ContainerDied","Data":"7c7d717b1e29083a433e71613e0683a5404c9b376452940760f7722e834726ff"} Dec 03 17:40:09 crc kubenswrapper[4998]: I1203 17:40:09.157012 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mqh5v" event={"ID":"4eed4b9a-97db-42f1-b43b-1b86dcb28b80","Type":"ContainerStarted","Data":"880997e33306fc19e5d5984530ab819a341d28d35045a432df01c71ae6ad1731"} Dec 03 17:40:10 crc kubenswrapper[4998]: I1203 17:40:10.167884 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mqh5v" event={"ID":"4eed4b9a-97db-42f1-b43b-1b86dcb28b80","Type":"ContainerStarted","Data":"860280a2225d4ab31ba7d1d05349206c0f6479a01e7da2c97da56330502c002b"} Dec 03 17:40:11 crc kubenswrapper[4998]: I1203 17:40:11.190289 4998 generic.go:334] "Generic (PLEG): container finished" podID="4eed4b9a-97db-42f1-b43b-1b86dcb28b80" containerID="860280a2225d4ab31ba7d1d05349206c0f6479a01e7da2c97da56330502c002b" exitCode=0 Dec 03 17:40:11 crc kubenswrapper[4998]: I1203 17:40:11.190409 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mqh5v" event={"ID":"4eed4b9a-97db-42f1-b43b-1b86dcb28b80","Type":"ContainerDied","Data":"860280a2225d4ab31ba7d1d05349206c0f6479a01e7da2c97da56330502c002b"} Dec 03 17:40:11 crc kubenswrapper[4998]: I1203 17:40:11.684732 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:40:11 crc kubenswrapper[4998]: E1203 17:40:11.685877 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:40:12 crc kubenswrapper[4998]: I1203 17:40:12.208721 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mqh5v" event={"ID":"4eed4b9a-97db-42f1-b43b-1b86dcb28b80","Type":"ContainerStarted","Data":"713870b960eed30b68ad1f1b26b8be535d97a2bf63192376a07cd4560b924b30"} Dec 03 17:40:12 crc kubenswrapper[4998]: I1203 17:40:12.245747 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mqh5v" podStartSLOduration=2.780002648 podStartE2EDuration="5.245728575s" podCreationTimestamp="2025-12-03 17:40:07 +0000 UTC" firstStartedPulling="2025-12-03 17:40:09.159658793 +0000 UTC m=+5787.771359036" lastFinishedPulling="2025-12-03 17:40:11.62538473 +0000 UTC m=+5790.237084963" observedRunningTime="2025-12-03 17:40:12.238668199 +0000 UTC m=+5790.850368472" watchObservedRunningTime="2025-12-03 17:40:12.245728575 +0000 UTC m=+5790.857428788" Dec 03 17:40:18 crc kubenswrapper[4998]: I1203 17:40:18.180019 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mqh5v" Dec 03 17:40:18 crc kubenswrapper[4998]: I1203 17:40:18.182037 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mqh5v" Dec 03 17:40:18 crc kubenswrapper[4998]: I1203 17:40:18.267667 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mqh5v" Dec 03 17:40:18 crc kubenswrapper[4998]: I1203 17:40:18.322489 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mqh5v" Dec 03 17:40:18 crc kubenswrapper[4998]: I1203 17:40:18.513455 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mqh5v"] Dec 03 17:40:20 crc kubenswrapper[4998]: I1203 17:40:20.289053 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mqh5v" podUID="4eed4b9a-97db-42f1-b43b-1b86dcb28b80" containerName="registry-server" containerID="cri-o://713870b960eed30b68ad1f1b26b8be535d97a2bf63192376a07cd4560b924b30" gracePeriod=2 Dec 03 17:40:21 crc kubenswrapper[4998]: I1203 17:40:21.308553 4998 generic.go:334] "Generic (PLEG): container finished" podID="4eed4b9a-97db-42f1-b43b-1b86dcb28b80" containerID="713870b960eed30b68ad1f1b26b8be535d97a2bf63192376a07cd4560b924b30" exitCode=0 Dec 03 17:40:21 crc kubenswrapper[4998]: I1203 17:40:21.309219 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mqh5v" event={"ID":"4eed4b9a-97db-42f1-b43b-1b86dcb28b80","Type":"ContainerDied","Data":"713870b960eed30b68ad1f1b26b8be535d97a2bf63192376a07cd4560b924b30"} Dec 03 17:40:21 crc kubenswrapper[4998]: I1203 17:40:21.309260 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mqh5v" event={"ID":"4eed4b9a-97db-42f1-b43b-1b86dcb28b80","Type":"ContainerDied","Data":"880997e33306fc19e5d5984530ab819a341d28d35045a432df01c71ae6ad1731"} Dec 03 17:40:21 crc kubenswrapper[4998]: I1203 17:40:21.309279 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="880997e33306fc19e5d5984530ab819a341d28d35045a432df01c71ae6ad1731" Dec 03 17:40:21 crc kubenswrapper[4998]: I1203 17:40:21.325824 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mqh5v" Dec 03 17:40:21 crc kubenswrapper[4998]: I1203 17:40:21.491176 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4eed4b9a-97db-42f1-b43b-1b86dcb28b80-utilities\") pod \"4eed4b9a-97db-42f1-b43b-1b86dcb28b80\" (UID: \"4eed4b9a-97db-42f1-b43b-1b86dcb28b80\") " Dec 03 17:40:21 crc kubenswrapper[4998]: I1203 17:40:21.492479 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4eed4b9a-97db-42f1-b43b-1b86dcb28b80-utilities" (OuterVolumeSpecName: "utilities") pod "4eed4b9a-97db-42f1-b43b-1b86dcb28b80" (UID: "4eed4b9a-97db-42f1-b43b-1b86dcb28b80"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:40:21 crc kubenswrapper[4998]: I1203 17:40:21.492651 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4eed4b9a-97db-42f1-b43b-1b86dcb28b80-catalog-content\") pod \"4eed4b9a-97db-42f1-b43b-1b86dcb28b80\" (UID: \"4eed4b9a-97db-42f1-b43b-1b86dcb28b80\") " Dec 03 17:40:21 crc kubenswrapper[4998]: I1203 17:40:21.495090 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdvwd\" (UniqueName: \"kubernetes.io/projected/4eed4b9a-97db-42f1-b43b-1b86dcb28b80-kube-api-access-wdvwd\") pod \"4eed4b9a-97db-42f1-b43b-1b86dcb28b80\" (UID: \"4eed4b9a-97db-42f1-b43b-1b86dcb28b80\") " Dec 03 17:40:21 crc kubenswrapper[4998]: I1203 17:40:21.496162 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4eed4b9a-97db-42f1-b43b-1b86dcb28b80-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:21 crc kubenswrapper[4998]: I1203 17:40:21.502826 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eed4b9a-97db-42f1-b43b-1b86dcb28b80-kube-api-access-wdvwd" (OuterVolumeSpecName: "kube-api-access-wdvwd") pod "4eed4b9a-97db-42f1-b43b-1b86dcb28b80" (UID: "4eed4b9a-97db-42f1-b43b-1b86dcb28b80"). InnerVolumeSpecName "kube-api-access-wdvwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:40:21 crc kubenswrapper[4998]: I1203 17:40:21.554908 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4eed4b9a-97db-42f1-b43b-1b86dcb28b80-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4eed4b9a-97db-42f1-b43b-1b86dcb28b80" (UID: "4eed4b9a-97db-42f1-b43b-1b86dcb28b80"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:40:21 crc kubenswrapper[4998]: I1203 17:40:21.597470 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4eed4b9a-97db-42f1-b43b-1b86dcb28b80-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:21 crc kubenswrapper[4998]: I1203 17:40:21.597508 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdvwd\" (UniqueName: \"kubernetes.io/projected/4eed4b9a-97db-42f1-b43b-1b86dcb28b80-kube-api-access-wdvwd\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:22 crc kubenswrapper[4998]: I1203 17:40:22.318009 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mqh5v" Dec 03 17:40:22 crc kubenswrapper[4998]: I1203 17:40:22.349398 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mqh5v"] Dec 03 17:40:22 crc kubenswrapper[4998]: I1203 17:40:22.360209 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mqh5v"] Dec 03 17:40:23 crc kubenswrapper[4998]: I1203 17:40:23.692044 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4eed4b9a-97db-42f1-b43b-1b86dcb28b80" path="/var/lib/kubelet/pods/4eed4b9a-97db-42f1-b43b-1b86dcb28b80/volumes" Dec 03 17:40:24 crc kubenswrapper[4998]: I1203 17:40:24.678337 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:40:24 crc kubenswrapper[4998]: E1203 17:40:24.678675 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:40:31 crc kubenswrapper[4998]: I1203 17:40:31.975082 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-l56fr"] Dec 03 17:40:31 crc kubenswrapper[4998]: E1203 17:40:31.977209 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eed4b9a-97db-42f1-b43b-1b86dcb28b80" containerName="extract-utilities" Dec 03 17:40:31 crc kubenswrapper[4998]: I1203 17:40:31.977244 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eed4b9a-97db-42f1-b43b-1b86dcb28b80" containerName="extract-utilities" Dec 03 17:40:31 crc kubenswrapper[4998]: E1203 17:40:31.977296 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eed4b9a-97db-42f1-b43b-1b86dcb28b80" containerName="registry-server" Dec 03 17:40:31 crc kubenswrapper[4998]: I1203 17:40:31.977464 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eed4b9a-97db-42f1-b43b-1b86dcb28b80" containerName="registry-server" Dec 03 17:40:31 crc kubenswrapper[4998]: E1203 17:40:31.977510 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eed4b9a-97db-42f1-b43b-1b86dcb28b80" containerName="extract-content" Dec 03 17:40:31 crc kubenswrapper[4998]: I1203 17:40:31.977527 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eed4b9a-97db-42f1-b43b-1b86dcb28b80" containerName="extract-content" Dec 03 17:40:31 crc kubenswrapper[4998]: I1203 17:40:31.978242 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eed4b9a-97db-42f1-b43b-1b86dcb28b80" containerName="registry-server" Dec 03 17:40:31 crc kubenswrapper[4998]: I1203 17:40:31.983577 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l56fr" Dec 03 17:40:32 crc kubenswrapper[4998]: I1203 17:40:32.022126 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l56fr"] Dec 03 17:40:32 crc kubenswrapper[4998]: I1203 17:40:32.028254 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6bf2dbf-9af1-4383-b031-9eb7f585685d-utilities\") pod \"certified-operators-l56fr\" (UID: \"f6bf2dbf-9af1-4383-b031-9eb7f585685d\") " pod="openshift-marketplace/certified-operators-l56fr" Dec 03 17:40:32 crc kubenswrapper[4998]: I1203 17:40:32.028329 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6bf2dbf-9af1-4383-b031-9eb7f585685d-catalog-content\") pod \"certified-operators-l56fr\" (UID: \"f6bf2dbf-9af1-4383-b031-9eb7f585685d\") " pod="openshift-marketplace/certified-operators-l56fr" Dec 03 17:40:32 crc kubenswrapper[4998]: I1203 17:40:32.028487 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fb4w\" (UniqueName: \"kubernetes.io/projected/f6bf2dbf-9af1-4383-b031-9eb7f585685d-kube-api-access-9fb4w\") pod \"certified-operators-l56fr\" (UID: \"f6bf2dbf-9af1-4383-b031-9eb7f585685d\") " pod="openshift-marketplace/certified-operators-l56fr" Dec 03 17:40:32 crc kubenswrapper[4998]: I1203 17:40:32.130408 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fb4w\" (UniqueName: \"kubernetes.io/projected/f6bf2dbf-9af1-4383-b031-9eb7f585685d-kube-api-access-9fb4w\") pod \"certified-operators-l56fr\" (UID: \"f6bf2dbf-9af1-4383-b031-9eb7f585685d\") " pod="openshift-marketplace/certified-operators-l56fr" Dec 03 17:40:32 crc kubenswrapper[4998]: I1203 17:40:32.130606 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6bf2dbf-9af1-4383-b031-9eb7f585685d-utilities\") pod \"certified-operators-l56fr\" (UID: \"f6bf2dbf-9af1-4383-b031-9eb7f585685d\") " pod="openshift-marketplace/certified-operators-l56fr" Dec 03 17:40:32 crc kubenswrapper[4998]: I1203 17:40:32.130635 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6bf2dbf-9af1-4383-b031-9eb7f585685d-catalog-content\") pod \"certified-operators-l56fr\" (UID: \"f6bf2dbf-9af1-4383-b031-9eb7f585685d\") " pod="openshift-marketplace/certified-operators-l56fr" Dec 03 17:40:32 crc kubenswrapper[4998]: I1203 17:40:32.131180 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6bf2dbf-9af1-4383-b031-9eb7f585685d-catalog-content\") pod \"certified-operators-l56fr\" (UID: \"f6bf2dbf-9af1-4383-b031-9eb7f585685d\") " pod="openshift-marketplace/certified-operators-l56fr" Dec 03 17:40:32 crc kubenswrapper[4998]: I1203 17:40:32.131212 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6bf2dbf-9af1-4383-b031-9eb7f585685d-utilities\") pod \"certified-operators-l56fr\" (UID: \"f6bf2dbf-9af1-4383-b031-9eb7f585685d\") " pod="openshift-marketplace/certified-operators-l56fr" Dec 03 17:40:32 crc kubenswrapper[4998]: I1203 17:40:32.161746 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fb4w\" (UniqueName: \"kubernetes.io/projected/f6bf2dbf-9af1-4383-b031-9eb7f585685d-kube-api-access-9fb4w\") pod \"certified-operators-l56fr\" (UID: \"f6bf2dbf-9af1-4383-b031-9eb7f585685d\") " pod="openshift-marketplace/certified-operators-l56fr" Dec 03 17:40:32 crc kubenswrapper[4998]: I1203 17:40:32.319646 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l56fr" Dec 03 17:40:32 crc kubenswrapper[4998]: I1203 17:40:32.945444 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l56fr"] Dec 03 17:40:33 crc kubenswrapper[4998]: I1203 17:40:33.453644 4998 generic.go:334] "Generic (PLEG): container finished" podID="f6bf2dbf-9af1-4383-b031-9eb7f585685d" containerID="6842445e8f0a154dead30340f44deaa7c5da0e6ea1b81e026949928ff4906125" exitCode=0 Dec 03 17:40:33 crc kubenswrapper[4998]: I1203 17:40:33.453768 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l56fr" event={"ID":"f6bf2dbf-9af1-4383-b031-9eb7f585685d","Type":"ContainerDied","Data":"6842445e8f0a154dead30340f44deaa7c5da0e6ea1b81e026949928ff4906125"} Dec 03 17:40:33 crc kubenswrapper[4998]: I1203 17:40:33.453890 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l56fr" event={"ID":"f6bf2dbf-9af1-4383-b031-9eb7f585685d","Type":"ContainerStarted","Data":"36d4bb7c674a9ed447835da8a1da71f7af2d9a1d46bef6b7e7e83edcc0a60171"} Dec 03 17:40:33 crc kubenswrapper[4998]: I1203 17:40:33.456584 4998 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 17:40:34 crc kubenswrapper[4998]: I1203 17:40:34.468827 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l56fr" event={"ID":"f6bf2dbf-9af1-4383-b031-9eb7f585685d","Type":"ContainerStarted","Data":"baeff9ed7221ac9c07fc9215a0b2b9185a3edc244467c726ebb1d82f9c7b20ef"} Dec 03 17:40:35 crc kubenswrapper[4998]: I1203 17:40:35.502815 4998 generic.go:334] "Generic (PLEG): container finished" podID="f6bf2dbf-9af1-4383-b031-9eb7f585685d" containerID="baeff9ed7221ac9c07fc9215a0b2b9185a3edc244467c726ebb1d82f9c7b20ef" exitCode=0 Dec 03 17:40:35 crc kubenswrapper[4998]: I1203 17:40:35.503060 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l56fr" event={"ID":"f6bf2dbf-9af1-4383-b031-9eb7f585685d","Type":"ContainerDied","Data":"baeff9ed7221ac9c07fc9215a0b2b9185a3edc244467c726ebb1d82f9c7b20ef"} Dec 03 17:40:35 crc kubenswrapper[4998]: I1203 17:40:35.678455 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:40:35 crc kubenswrapper[4998]: E1203 17:40:35.678735 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:40:36 crc kubenswrapper[4998]: I1203 17:40:36.514703 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l56fr" event={"ID":"f6bf2dbf-9af1-4383-b031-9eb7f585685d","Type":"ContainerStarted","Data":"10c432e9bf5818817c22cf091e05588ea27df9e059343d5e1d9fbb2d28e29ca5"} Dec 03 17:40:36 crc kubenswrapper[4998]: I1203 17:40:36.540453 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-l56fr" podStartSLOduration=2.865525932 podStartE2EDuration="5.540432754s" podCreationTimestamp="2025-12-03 17:40:31 +0000 UTC" firstStartedPulling="2025-12-03 17:40:33.456362361 +0000 UTC m=+5812.068062584" lastFinishedPulling="2025-12-03 17:40:36.131269173 +0000 UTC m=+5814.742969406" observedRunningTime="2025-12-03 17:40:36.538110626 +0000 UTC m=+5815.149810849" watchObservedRunningTime="2025-12-03 17:40:36.540432754 +0000 UTC m=+5815.152132987" Dec 03 17:40:42 crc kubenswrapper[4998]: I1203 17:40:42.320147 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-l56fr" Dec 03 17:40:42 crc kubenswrapper[4998]: I1203 17:40:42.321403 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-l56fr" Dec 03 17:40:42 crc kubenswrapper[4998]: I1203 17:40:42.371860 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-l56fr" Dec 03 17:40:42 crc kubenswrapper[4998]: I1203 17:40:42.631956 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-l56fr" Dec 03 17:40:42 crc kubenswrapper[4998]: I1203 17:40:42.692215 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l56fr"] Dec 03 17:40:44 crc kubenswrapper[4998]: I1203 17:40:44.626319 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-l56fr" podUID="f6bf2dbf-9af1-4383-b031-9eb7f585685d" containerName="registry-server" containerID="cri-o://10c432e9bf5818817c22cf091e05588ea27df9e059343d5e1d9fbb2d28e29ca5" gracePeriod=2 Dec 03 17:40:45 crc kubenswrapper[4998]: I1203 17:40:45.154230 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l56fr" Dec 03 17:40:45 crc kubenswrapper[4998]: I1203 17:40:45.246398 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6bf2dbf-9af1-4383-b031-9eb7f585685d-catalog-content\") pod \"f6bf2dbf-9af1-4383-b031-9eb7f585685d\" (UID: \"f6bf2dbf-9af1-4383-b031-9eb7f585685d\") " Dec 03 17:40:45 crc kubenswrapper[4998]: I1203 17:40:45.246505 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fb4w\" (UniqueName: \"kubernetes.io/projected/f6bf2dbf-9af1-4383-b031-9eb7f585685d-kube-api-access-9fb4w\") pod \"f6bf2dbf-9af1-4383-b031-9eb7f585685d\" (UID: \"f6bf2dbf-9af1-4383-b031-9eb7f585685d\") " Dec 03 17:40:45 crc kubenswrapper[4998]: I1203 17:40:45.246562 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6bf2dbf-9af1-4383-b031-9eb7f585685d-utilities\") pod \"f6bf2dbf-9af1-4383-b031-9eb7f585685d\" (UID: \"f6bf2dbf-9af1-4383-b031-9eb7f585685d\") " Dec 03 17:40:45 crc kubenswrapper[4998]: I1203 17:40:45.247535 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6bf2dbf-9af1-4383-b031-9eb7f585685d-utilities" (OuterVolumeSpecName: "utilities") pod "f6bf2dbf-9af1-4383-b031-9eb7f585685d" (UID: "f6bf2dbf-9af1-4383-b031-9eb7f585685d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:40:45 crc kubenswrapper[4998]: I1203 17:40:45.252379 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6bf2dbf-9af1-4383-b031-9eb7f585685d-kube-api-access-9fb4w" (OuterVolumeSpecName: "kube-api-access-9fb4w") pod "f6bf2dbf-9af1-4383-b031-9eb7f585685d" (UID: "f6bf2dbf-9af1-4383-b031-9eb7f585685d"). InnerVolumeSpecName "kube-api-access-9fb4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:40:45 crc kubenswrapper[4998]: I1203 17:40:45.293980 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6bf2dbf-9af1-4383-b031-9eb7f585685d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f6bf2dbf-9af1-4383-b031-9eb7f585685d" (UID: "f6bf2dbf-9af1-4383-b031-9eb7f585685d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:40:45 crc kubenswrapper[4998]: I1203 17:40:45.349152 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fb4w\" (UniqueName: \"kubernetes.io/projected/f6bf2dbf-9af1-4383-b031-9eb7f585685d-kube-api-access-9fb4w\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:45 crc kubenswrapper[4998]: I1203 17:40:45.349193 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6bf2dbf-9af1-4383-b031-9eb7f585685d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:45 crc kubenswrapper[4998]: I1203 17:40:45.349212 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6bf2dbf-9af1-4383-b031-9eb7f585685d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:45 crc kubenswrapper[4998]: I1203 17:40:45.637997 4998 generic.go:334] "Generic (PLEG): container finished" podID="f6bf2dbf-9af1-4383-b031-9eb7f585685d" containerID="10c432e9bf5818817c22cf091e05588ea27df9e059343d5e1d9fbb2d28e29ca5" exitCode=0 Dec 03 17:40:45 crc kubenswrapper[4998]: I1203 17:40:45.638051 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l56fr" Dec 03 17:40:45 crc kubenswrapper[4998]: I1203 17:40:45.638049 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l56fr" event={"ID":"f6bf2dbf-9af1-4383-b031-9eb7f585685d","Type":"ContainerDied","Data":"10c432e9bf5818817c22cf091e05588ea27df9e059343d5e1d9fbb2d28e29ca5"} Dec 03 17:40:45 crc kubenswrapper[4998]: I1203 17:40:45.638311 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l56fr" event={"ID":"f6bf2dbf-9af1-4383-b031-9eb7f585685d","Type":"ContainerDied","Data":"36d4bb7c674a9ed447835da8a1da71f7af2d9a1d46bef6b7e7e83edcc0a60171"} Dec 03 17:40:45 crc kubenswrapper[4998]: I1203 17:40:45.638353 4998 scope.go:117] "RemoveContainer" containerID="10c432e9bf5818817c22cf091e05588ea27df9e059343d5e1d9fbb2d28e29ca5" Dec 03 17:40:45 crc kubenswrapper[4998]: I1203 17:40:45.674211 4998 scope.go:117] "RemoveContainer" containerID="baeff9ed7221ac9c07fc9215a0b2b9185a3edc244467c726ebb1d82f9c7b20ef" Dec 03 17:40:45 crc kubenswrapper[4998]: I1203 17:40:45.701294 4998 scope.go:117] "RemoveContainer" containerID="6842445e8f0a154dead30340f44deaa7c5da0e6ea1b81e026949928ff4906125" Dec 03 17:40:45 crc kubenswrapper[4998]: I1203 17:40:45.703791 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l56fr"] Dec 03 17:40:45 crc kubenswrapper[4998]: I1203 17:40:45.703824 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-l56fr"] Dec 03 17:40:45 crc kubenswrapper[4998]: I1203 17:40:45.751911 4998 scope.go:117] "RemoveContainer" containerID="10c432e9bf5818817c22cf091e05588ea27df9e059343d5e1d9fbb2d28e29ca5" Dec 03 17:40:45 crc kubenswrapper[4998]: E1203 17:40:45.752365 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10c432e9bf5818817c22cf091e05588ea27df9e059343d5e1d9fbb2d28e29ca5\": container with ID starting with 10c432e9bf5818817c22cf091e05588ea27df9e059343d5e1d9fbb2d28e29ca5 not found: ID does not exist" containerID="10c432e9bf5818817c22cf091e05588ea27df9e059343d5e1d9fbb2d28e29ca5" Dec 03 17:40:45 crc kubenswrapper[4998]: I1203 17:40:45.752422 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10c432e9bf5818817c22cf091e05588ea27df9e059343d5e1d9fbb2d28e29ca5"} err="failed to get container status \"10c432e9bf5818817c22cf091e05588ea27df9e059343d5e1d9fbb2d28e29ca5\": rpc error: code = NotFound desc = could not find container \"10c432e9bf5818817c22cf091e05588ea27df9e059343d5e1d9fbb2d28e29ca5\": container with ID starting with 10c432e9bf5818817c22cf091e05588ea27df9e059343d5e1d9fbb2d28e29ca5 not found: ID does not exist" Dec 03 17:40:45 crc kubenswrapper[4998]: I1203 17:40:45.752455 4998 scope.go:117] "RemoveContainer" containerID="baeff9ed7221ac9c07fc9215a0b2b9185a3edc244467c726ebb1d82f9c7b20ef" Dec 03 17:40:45 crc kubenswrapper[4998]: E1203 17:40:45.752812 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"baeff9ed7221ac9c07fc9215a0b2b9185a3edc244467c726ebb1d82f9c7b20ef\": container with ID starting with baeff9ed7221ac9c07fc9215a0b2b9185a3edc244467c726ebb1d82f9c7b20ef not found: ID does not exist" containerID="baeff9ed7221ac9c07fc9215a0b2b9185a3edc244467c726ebb1d82f9c7b20ef" Dec 03 17:40:45 crc kubenswrapper[4998]: I1203 17:40:45.752845 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baeff9ed7221ac9c07fc9215a0b2b9185a3edc244467c726ebb1d82f9c7b20ef"} err="failed to get container status \"baeff9ed7221ac9c07fc9215a0b2b9185a3edc244467c726ebb1d82f9c7b20ef\": rpc error: code = NotFound desc = could not find container \"baeff9ed7221ac9c07fc9215a0b2b9185a3edc244467c726ebb1d82f9c7b20ef\": container with ID starting with baeff9ed7221ac9c07fc9215a0b2b9185a3edc244467c726ebb1d82f9c7b20ef not found: ID does not exist" Dec 03 17:40:45 crc kubenswrapper[4998]: I1203 17:40:45.752860 4998 scope.go:117] "RemoveContainer" containerID="6842445e8f0a154dead30340f44deaa7c5da0e6ea1b81e026949928ff4906125" Dec 03 17:40:45 crc kubenswrapper[4998]: E1203 17:40:45.753269 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6842445e8f0a154dead30340f44deaa7c5da0e6ea1b81e026949928ff4906125\": container with ID starting with 6842445e8f0a154dead30340f44deaa7c5da0e6ea1b81e026949928ff4906125 not found: ID does not exist" containerID="6842445e8f0a154dead30340f44deaa7c5da0e6ea1b81e026949928ff4906125" Dec 03 17:40:45 crc kubenswrapper[4998]: I1203 17:40:45.753287 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6842445e8f0a154dead30340f44deaa7c5da0e6ea1b81e026949928ff4906125"} err="failed to get container status \"6842445e8f0a154dead30340f44deaa7c5da0e6ea1b81e026949928ff4906125\": rpc error: code = NotFound desc = could not find container \"6842445e8f0a154dead30340f44deaa7c5da0e6ea1b81e026949928ff4906125\": container with ID starting with 6842445e8f0a154dead30340f44deaa7c5da0e6ea1b81e026949928ff4906125 not found: ID does not exist" Dec 03 17:40:46 crc kubenswrapper[4998]: I1203 17:40:46.679087 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:40:46 crc kubenswrapper[4998]: E1203 17:40:46.679933 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:40:47 crc kubenswrapper[4998]: I1203 17:40:47.691010 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6bf2dbf-9af1-4383-b031-9eb7f585685d" path="/var/lib/kubelet/pods/f6bf2dbf-9af1-4383-b031-9eb7f585685d/volumes" Dec 03 17:40:59 crc kubenswrapper[4998]: I1203 17:40:59.681397 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:40:59 crc kubenswrapper[4998]: E1203 17:40:59.683219 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:41:12 crc kubenswrapper[4998]: I1203 17:41:12.678611 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:41:12 crc kubenswrapper[4998]: E1203 17:41:12.679515 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:41:26 crc kubenswrapper[4998]: I1203 17:41:26.678175 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:41:26 crc kubenswrapper[4998]: E1203 17:41:26.679253 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:41:41 crc kubenswrapper[4998]: I1203 17:41:41.684542 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:41:41 crc kubenswrapper[4998]: E1203 17:41:41.685227 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:41:52 crc kubenswrapper[4998]: I1203 17:41:52.678914 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:41:52 crc kubenswrapper[4998]: E1203 17:41:52.679681 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:42:02 crc kubenswrapper[4998]: I1203 17:42:02.976978 4998 scope.go:117] "RemoveContainer" containerID="ebfdd11f4088e9b04bede7901d43048054fa260a26bbcebacbed7ffa0dd1b25f" Dec 03 17:42:03 crc kubenswrapper[4998]: I1203 17:42:03.027939 4998 scope.go:117] "RemoveContainer" containerID="c9c85fcf851bc29fc2f9fb30304caa56022524b3943dc2be71b048e740e96d92" Dec 03 17:42:03 crc kubenswrapper[4998]: I1203 17:42:03.088812 4998 scope.go:117] "RemoveContainer" containerID="2ddc951b7a87bccf82e70a6c6d5e3a2e9ff92f20d5717fe686efddf0c1d8229d" Dec 03 17:42:04 crc kubenswrapper[4998]: I1203 17:42:04.677848 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:42:05 crc kubenswrapper[4998]: I1203 17:42:05.479795 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"c32ad81ca8fbe8808005dc071a77b0419d9a93bdb054bce341dafb8dbf1dd822"} Dec 03 17:43:40 crc kubenswrapper[4998]: I1203 17:43:40.571687 4998 generic.go:334] "Generic (PLEG): container finished" podID="27edc0ce-edf8-4c0e-bd5f-79e58650eed1" containerID="55cea1c7702ffa07417318378686a0b88bb4d847c19bf491fc95f56a149afbc4" exitCode=1 Dec 03 17:43:40 crc kubenswrapper[4998]: I1203 17:43:40.571833 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"27edc0ce-edf8-4c0e-bd5f-79e58650eed1","Type":"ContainerDied","Data":"55cea1c7702ffa07417318378686a0b88bb4d847c19bf491fc95f56a149afbc4"} Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.014971 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.069700 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-config-data\") pod \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.069853 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-openstack-config\") pod \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.070139 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-test-operator-ephemeral-workdir\") pod \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.070239 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-ssh-key\") pod \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.070339 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-ca-certs\") pod \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.070479 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-openstack-config-secret\") pod \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.070565 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.070625 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-test-operator-ephemeral-temporary\") pod \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.070689 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ww9s\" (UniqueName: \"kubernetes.io/projected/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-kube-api-access-8ww9s\") pod \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\" (UID: \"27edc0ce-edf8-4c0e-bd5f-79e58650eed1\") " Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.073968 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "27edc0ce-edf8-4c0e-bd5f-79e58650eed1" (UID: "27edc0ce-edf8-4c0e-bd5f-79e58650eed1"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.074288 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-config-data" (OuterVolumeSpecName: "config-data") pod "27edc0ce-edf8-4c0e-bd5f-79e58650eed1" (UID: "27edc0ce-edf8-4c0e-bd5f-79e58650eed1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.083995 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-kube-api-access-8ww9s" (OuterVolumeSpecName: "kube-api-access-8ww9s") pod "27edc0ce-edf8-4c0e-bd5f-79e58650eed1" (UID: "27edc0ce-edf8-4c0e-bd5f-79e58650eed1"). InnerVolumeSpecName "kube-api-access-8ww9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.085077 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "test-operator-logs") pod "27edc0ce-edf8-4c0e-bd5f-79e58650eed1" (UID: "27edc0ce-edf8-4c0e-bd5f-79e58650eed1"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.093954 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "27edc0ce-edf8-4c0e-bd5f-79e58650eed1" (UID: "27edc0ce-edf8-4c0e-bd5f-79e58650eed1"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.126106 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "27edc0ce-edf8-4c0e-bd5f-79e58650eed1" (UID: "27edc0ce-edf8-4c0e-bd5f-79e58650eed1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.126137 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "27edc0ce-edf8-4c0e-bd5f-79e58650eed1" (UID: "27edc0ce-edf8-4c0e-bd5f-79e58650eed1"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.129032 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "27edc0ce-edf8-4c0e-bd5f-79e58650eed1" (UID: "27edc0ce-edf8-4c0e-bd5f-79e58650eed1"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.131936 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "27edc0ce-edf8-4c0e-bd5f-79e58650eed1" (UID: "27edc0ce-edf8-4c0e-bd5f-79e58650eed1"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.174525 4998 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.175129 4998 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.175230 4998 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.176083 4998 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.176222 4998 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.176298 4998 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.176364 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ww9s\" (UniqueName: \"kubernetes.io/projected/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-kube-api-access-8ww9s\") on node \"crc\" DevicePath \"\"" Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.176428 4998 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.176491 4998 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/27edc0ce-edf8-4c0e-bd5f-79e58650eed1-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.196839 4998 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.279446 4998 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.601428 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"27edc0ce-edf8-4c0e-bd5f-79e58650eed1","Type":"ContainerDied","Data":"63a2fe08a9bdff7f5ed4f9daa8d5547870580605d42e0286337953d2a8d6b65b"} Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.601497 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63a2fe08a9bdff7f5ed4f9daa8d5547870580605d42e0286337953d2a8d6b65b" Dec 03 17:43:42 crc kubenswrapper[4998]: I1203 17:43:42.601511 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 17:43:45 crc kubenswrapper[4998]: I1203 17:43:45.073419 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 03 17:43:45 crc kubenswrapper[4998]: E1203 17:43:45.074240 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6bf2dbf-9af1-4383-b031-9eb7f585685d" containerName="extract-utilities" Dec 03 17:43:45 crc kubenswrapper[4998]: I1203 17:43:45.074257 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6bf2dbf-9af1-4383-b031-9eb7f585685d" containerName="extract-utilities" Dec 03 17:43:45 crc kubenswrapper[4998]: E1203 17:43:45.074280 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6bf2dbf-9af1-4383-b031-9eb7f585685d" containerName="registry-server" Dec 03 17:43:45 crc kubenswrapper[4998]: I1203 17:43:45.074287 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6bf2dbf-9af1-4383-b031-9eb7f585685d" containerName="registry-server" Dec 03 17:43:45 crc kubenswrapper[4998]: E1203 17:43:45.074325 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27edc0ce-edf8-4c0e-bd5f-79e58650eed1" containerName="tempest-tests-tempest-tests-runner" Dec 03 17:43:45 crc kubenswrapper[4998]: I1203 17:43:45.074334 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="27edc0ce-edf8-4c0e-bd5f-79e58650eed1" containerName="tempest-tests-tempest-tests-runner" Dec 03 17:43:45 crc kubenswrapper[4998]: E1203 17:43:45.074359 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6bf2dbf-9af1-4383-b031-9eb7f585685d" containerName="extract-content" Dec 03 17:43:45 crc kubenswrapper[4998]: I1203 17:43:45.074369 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6bf2dbf-9af1-4383-b031-9eb7f585685d" containerName="extract-content" Dec 03 17:43:45 crc kubenswrapper[4998]: I1203 17:43:45.074596 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="27edc0ce-edf8-4c0e-bd5f-79e58650eed1" containerName="tempest-tests-tempest-tests-runner" Dec 03 17:43:45 crc kubenswrapper[4998]: I1203 17:43:45.074615 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6bf2dbf-9af1-4383-b031-9eb7f585685d" containerName="registry-server" Dec 03 17:43:45 crc kubenswrapper[4998]: I1203 17:43:45.075432 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 17:43:45 crc kubenswrapper[4998]: I1203 17:43:45.080859 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-28bzn" Dec 03 17:43:45 crc kubenswrapper[4998]: I1203 17:43:45.085141 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 03 17:43:45 crc kubenswrapper[4998]: I1203 17:43:45.142650 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"eb306f4f-0f10-48e1-b20d-78b444da8b8e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 17:43:45 crc kubenswrapper[4998]: I1203 17:43:45.142720 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7sms\" (UniqueName: \"kubernetes.io/projected/eb306f4f-0f10-48e1-b20d-78b444da8b8e-kube-api-access-f7sms\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"eb306f4f-0f10-48e1-b20d-78b444da8b8e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 17:43:45 crc kubenswrapper[4998]: I1203 17:43:45.244323 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"eb306f4f-0f10-48e1-b20d-78b444da8b8e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 17:43:45 crc kubenswrapper[4998]: I1203 17:43:45.244400 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7sms\" (UniqueName: \"kubernetes.io/projected/eb306f4f-0f10-48e1-b20d-78b444da8b8e-kube-api-access-f7sms\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"eb306f4f-0f10-48e1-b20d-78b444da8b8e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 17:43:45 crc kubenswrapper[4998]: I1203 17:43:45.244746 4998 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"eb306f4f-0f10-48e1-b20d-78b444da8b8e\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 17:43:45 crc kubenswrapper[4998]: I1203 17:43:45.271716 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7sms\" (UniqueName: \"kubernetes.io/projected/eb306f4f-0f10-48e1-b20d-78b444da8b8e-kube-api-access-f7sms\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"eb306f4f-0f10-48e1-b20d-78b444da8b8e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 17:43:45 crc kubenswrapper[4998]: I1203 17:43:45.289047 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"eb306f4f-0f10-48e1-b20d-78b444da8b8e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 17:43:45 crc kubenswrapper[4998]: I1203 17:43:45.416164 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 17:43:45 crc kubenswrapper[4998]: I1203 17:43:45.902180 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 03 17:43:46 crc kubenswrapper[4998]: I1203 17:43:46.648414 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"eb306f4f-0f10-48e1-b20d-78b444da8b8e","Type":"ContainerStarted","Data":"d1b06cd54fdb0a4355a78e2cb0132991b7cd58769c639e7c87c90878e087ae3a"} Dec 03 17:43:47 crc kubenswrapper[4998]: I1203 17:43:47.664875 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"eb306f4f-0f10-48e1-b20d-78b444da8b8e","Type":"ContainerStarted","Data":"9a7198903b2a45ce4ad3103079a1d801cb7a832194bf626262eeecee9a479d37"} Dec 03 17:43:47 crc kubenswrapper[4998]: I1203 17:43:47.693183 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.931412168 podStartE2EDuration="2.693156471s" podCreationTimestamp="2025-12-03 17:43:45 +0000 UTC" firstStartedPulling="2025-12-03 17:43:45.899000042 +0000 UTC m=+6004.510700265" lastFinishedPulling="2025-12-03 17:43:46.660744345 +0000 UTC m=+6005.272444568" observedRunningTime="2025-12-03 17:43:47.690812852 +0000 UTC m=+6006.302513145" watchObservedRunningTime="2025-12-03 17:43:47.693156471 +0000 UTC m=+6006.304856724" Dec 03 17:44:17 crc kubenswrapper[4998]: I1203 17:44:17.900798 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jx2sk/must-gather-nbb85"] Dec 03 17:44:17 crc kubenswrapper[4998]: I1203 17:44:17.902715 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jx2sk/must-gather-nbb85" Dec 03 17:44:17 crc kubenswrapper[4998]: I1203 17:44:17.905440 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-jx2sk"/"default-dockercfg-hm2g5" Dec 03 17:44:17 crc kubenswrapper[4998]: I1203 17:44:17.905885 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-jx2sk"/"kube-root-ca.crt" Dec 03 17:44:17 crc kubenswrapper[4998]: I1203 17:44:17.905970 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-jx2sk"/"openshift-service-ca.crt" Dec 03 17:44:17 crc kubenswrapper[4998]: I1203 17:44:17.910451 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jx2sk/must-gather-nbb85"] Dec 03 17:44:18 crc kubenswrapper[4998]: I1203 17:44:18.020739 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vh92\" (UniqueName: \"kubernetes.io/projected/5f4e13e1-8408-4046-a8b4-7591e8a1bd22-kube-api-access-7vh92\") pod \"must-gather-nbb85\" (UID: \"5f4e13e1-8408-4046-a8b4-7591e8a1bd22\") " pod="openshift-must-gather-jx2sk/must-gather-nbb85" Dec 03 17:44:18 crc kubenswrapper[4998]: I1203 17:44:18.021264 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5f4e13e1-8408-4046-a8b4-7591e8a1bd22-must-gather-output\") pod \"must-gather-nbb85\" (UID: \"5f4e13e1-8408-4046-a8b4-7591e8a1bd22\") " pod="openshift-must-gather-jx2sk/must-gather-nbb85" Dec 03 17:44:18 crc kubenswrapper[4998]: I1203 17:44:18.122879 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vh92\" (UniqueName: \"kubernetes.io/projected/5f4e13e1-8408-4046-a8b4-7591e8a1bd22-kube-api-access-7vh92\") pod \"must-gather-nbb85\" (UID: \"5f4e13e1-8408-4046-a8b4-7591e8a1bd22\") " pod="openshift-must-gather-jx2sk/must-gather-nbb85" Dec 03 17:44:18 crc kubenswrapper[4998]: I1203 17:44:18.122978 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5f4e13e1-8408-4046-a8b4-7591e8a1bd22-must-gather-output\") pod \"must-gather-nbb85\" (UID: \"5f4e13e1-8408-4046-a8b4-7591e8a1bd22\") " pod="openshift-must-gather-jx2sk/must-gather-nbb85" Dec 03 17:44:18 crc kubenswrapper[4998]: I1203 17:44:18.123555 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5f4e13e1-8408-4046-a8b4-7591e8a1bd22-must-gather-output\") pod \"must-gather-nbb85\" (UID: \"5f4e13e1-8408-4046-a8b4-7591e8a1bd22\") " pod="openshift-must-gather-jx2sk/must-gather-nbb85" Dec 03 17:44:18 crc kubenswrapper[4998]: I1203 17:44:18.141399 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vh92\" (UniqueName: \"kubernetes.io/projected/5f4e13e1-8408-4046-a8b4-7591e8a1bd22-kube-api-access-7vh92\") pod \"must-gather-nbb85\" (UID: \"5f4e13e1-8408-4046-a8b4-7591e8a1bd22\") " pod="openshift-must-gather-jx2sk/must-gather-nbb85" Dec 03 17:44:18 crc kubenswrapper[4998]: I1203 17:44:18.221697 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jx2sk/must-gather-nbb85" Dec 03 17:44:18 crc kubenswrapper[4998]: I1203 17:44:18.702321 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jx2sk/must-gather-nbb85"] Dec 03 17:44:19 crc kubenswrapper[4998]: I1203 17:44:19.061016 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jx2sk/must-gather-nbb85" event={"ID":"5f4e13e1-8408-4046-a8b4-7591e8a1bd22","Type":"ContainerStarted","Data":"cd62379692ce8b2052600e8bd2849b0119b7593f74107ac1e2bf8bc24c295846"} Dec 03 17:44:27 crc kubenswrapper[4998]: I1203 17:44:27.111739 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:44:27 crc kubenswrapper[4998]: I1203 17:44:27.112480 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:44:27 crc kubenswrapper[4998]: I1203 17:44:27.165242 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jx2sk/must-gather-nbb85" event={"ID":"5f4e13e1-8408-4046-a8b4-7591e8a1bd22","Type":"ContainerStarted","Data":"156ed544489d596d38766112ab42a92f8d9d6cfa55075b8d2e59bad36d942f78"} Dec 03 17:44:28 crc kubenswrapper[4998]: I1203 17:44:28.180934 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jx2sk/must-gather-nbb85" event={"ID":"5f4e13e1-8408-4046-a8b4-7591e8a1bd22","Type":"ContainerStarted","Data":"7615d5fcfd5c593a72110f7c17562e3cb44baf60e638c9f16434f81a459e543e"} Dec 03 17:44:28 crc kubenswrapper[4998]: I1203 17:44:28.206657 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jx2sk/must-gather-nbb85" podStartSLOduration=3.12222225 podStartE2EDuration="11.206641488s" podCreationTimestamp="2025-12-03 17:44:17 +0000 UTC" firstStartedPulling="2025-12-03 17:44:18.711478341 +0000 UTC m=+6037.323178564" lastFinishedPulling="2025-12-03 17:44:26.795897579 +0000 UTC m=+6045.407597802" observedRunningTime="2025-12-03 17:44:28.200901476 +0000 UTC m=+6046.812601699" watchObservedRunningTime="2025-12-03 17:44:28.206641488 +0000 UTC m=+6046.818341711" Dec 03 17:44:30 crc kubenswrapper[4998]: I1203 17:44:30.956840 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jx2sk/crc-debug-2rj8m"] Dec 03 17:44:30 crc kubenswrapper[4998]: I1203 17:44:30.958826 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jx2sk/crc-debug-2rj8m" Dec 03 17:44:31 crc kubenswrapper[4998]: I1203 17:44:31.037078 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/df3bf50a-6c2f-421e-8e35-3bcc2d368c6e-host\") pod \"crc-debug-2rj8m\" (UID: \"df3bf50a-6c2f-421e-8e35-3bcc2d368c6e\") " pod="openshift-must-gather-jx2sk/crc-debug-2rj8m" Dec 03 17:44:31 crc kubenswrapper[4998]: I1203 17:44:31.037275 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6t95\" (UniqueName: \"kubernetes.io/projected/df3bf50a-6c2f-421e-8e35-3bcc2d368c6e-kube-api-access-k6t95\") pod \"crc-debug-2rj8m\" (UID: \"df3bf50a-6c2f-421e-8e35-3bcc2d368c6e\") " pod="openshift-must-gather-jx2sk/crc-debug-2rj8m" Dec 03 17:44:31 crc kubenswrapper[4998]: I1203 17:44:31.139157 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6t95\" (UniqueName: \"kubernetes.io/projected/df3bf50a-6c2f-421e-8e35-3bcc2d368c6e-kube-api-access-k6t95\") pod \"crc-debug-2rj8m\" (UID: \"df3bf50a-6c2f-421e-8e35-3bcc2d368c6e\") " pod="openshift-must-gather-jx2sk/crc-debug-2rj8m" Dec 03 17:44:31 crc kubenswrapper[4998]: I1203 17:44:31.139255 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/df3bf50a-6c2f-421e-8e35-3bcc2d368c6e-host\") pod \"crc-debug-2rj8m\" (UID: \"df3bf50a-6c2f-421e-8e35-3bcc2d368c6e\") " pod="openshift-must-gather-jx2sk/crc-debug-2rj8m" Dec 03 17:44:31 crc kubenswrapper[4998]: I1203 17:44:31.139368 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/df3bf50a-6c2f-421e-8e35-3bcc2d368c6e-host\") pod \"crc-debug-2rj8m\" (UID: \"df3bf50a-6c2f-421e-8e35-3bcc2d368c6e\") " pod="openshift-must-gather-jx2sk/crc-debug-2rj8m" Dec 03 17:44:31 crc kubenswrapper[4998]: I1203 17:44:31.169655 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6t95\" (UniqueName: \"kubernetes.io/projected/df3bf50a-6c2f-421e-8e35-3bcc2d368c6e-kube-api-access-k6t95\") pod \"crc-debug-2rj8m\" (UID: \"df3bf50a-6c2f-421e-8e35-3bcc2d368c6e\") " pod="openshift-must-gather-jx2sk/crc-debug-2rj8m" Dec 03 17:44:31 crc kubenswrapper[4998]: I1203 17:44:31.291958 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jx2sk/crc-debug-2rj8m" Dec 03 17:44:32 crc kubenswrapper[4998]: I1203 17:44:32.221704 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jx2sk/crc-debug-2rj8m" event={"ID":"df3bf50a-6c2f-421e-8e35-3bcc2d368c6e","Type":"ContainerStarted","Data":"c8d5358ad0e14965fc79baa1749747a86d73b198ded3fd93d237168be1dc2aa7"} Dec 03 17:44:45 crc kubenswrapper[4998]: I1203 17:44:45.361155 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jx2sk/crc-debug-2rj8m" event={"ID":"df3bf50a-6c2f-421e-8e35-3bcc2d368c6e","Type":"ContainerStarted","Data":"043c895a485290d169a5f9c1316bc03c62bc355e855ead7fed014c4a3b5e296a"} Dec 03 17:44:45 crc kubenswrapper[4998]: I1203 17:44:45.375591 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jx2sk/crc-debug-2rj8m" podStartSLOduration=2.535684172 podStartE2EDuration="15.375570927s" podCreationTimestamp="2025-12-03 17:44:30 +0000 UTC" firstStartedPulling="2025-12-03 17:44:31.325180821 +0000 UTC m=+6049.936881034" lastFinishedPulling="2025-12-03 17:44:44.165067566 +0000 UTC m=+6062.776767789" observedRunningTime="2025-12-03 17:44:45.373598018 +0000 UTC m=+6063.985298291" watchObservedRunningTime="2025-12-03 17:44:45.375570927 +0000 UTC m=+6063.987271160" Dec 03 17:44:57 crc kubenswrapper[4998]: I1203 17:44:57.110856 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:44:57 crc kubenswrapper[4998]: I1203 17:44:57.111461 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:45:00 crc kubenswrapper[4998]: I1203 17:45:00.153186 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413065-8rf5w"] Dec 03 17:45:00 crc kubenswrapper[4998]: I1203 17:45:00.159479 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8rf5w" Dec 03 17:45:00 crc kubenswrapper[4998]: I1203 17:45:00.161537 4998 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 17:45:00 crc kubenswrapper[4998]: I1203 17:45:00.162265 4998 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 17:45:00 crc kubenswrapper[4998]: I1203 17:45:00.182300 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413065-8rf5w"] Dec 03 17:45:00 crc kubenswrapper[4998]: I1203 17:45:00.282724 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5-config-volume\") pod \"collect-profiles-29413065-8rf5w\" (UID: \"5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8rf5w" Dec 03 17:45:00 crc kubenswrapper[4998]: I1203 17:45:00.282799 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwffz\" (UniqueName: \"kubernetes.io/projected/5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5-kube-api-access-hwffz\") pod \"collect-profiles-29413065-8rf5w\" (UID: \"5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8rf5w" Dec 03 17:45:00 crc kubenswrapper[4998]: I1203 17:45:00.282874 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5-secret-volume\") pod \"collect-profiles-29413065-8rf5w\" (UID: \"5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8rf5w" Dec 03 17:45:00 crc kubenswrapper[4998]: I1203 17:45:00.384383 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwffz\" (UniqueName: \"kubernetes.io/projected/5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5-kube-api-access-hwffz\") pod \"collect-profiles-29413065-8rf5w\" (UID: \"5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8rf5w" Dec 03 17:45:00 crc kubenswrapper[4998]: I1203 17:45:00.384528 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5-secret-volume\") pod \"collect-profiles-29413065-8rf5w\" (UID: \"5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8rf5w" Dec 03 17:45:00 crc kubenswrapper[4998]: I1203 17:45:00.384707 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5-config-volume\") pod \"collect-profiles-29413065-8rf5w\" (UID: \"5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8rf5w" Dec 03 17:45:00 crc kubenswrapper[4998]: I1203 17:45:00.386063 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5-config-volume\") pod \"collect-profiles-29413065-8rf5w\" (UID: \"5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8rf5w" Dec 03 17:45:00 crc kubenswrapper[4998]: I1203 17:45:00.391745 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5-secret-volume\") pod \"collect-profiles-29413065-8rf5w\" (UID: \"5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8rf5w" Dec 03 17:45:00 crc kubenswrapper[4998]: I1203 17:45:00.399733 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwffz\" (UniqueName: \"kubernetes.io/projected/5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5-kube-api-access-hwffz\") pod \"collect-profiles-29413065-8rf5w\" (UID: \"5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8rf5w" Dec 03 17:45:00 crc kubenswrapper[4998]: I1203 17:45:00.498626 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8rf5w" Dec 03 17:45:00 crc kubenswrapper[4998]: W1203 17:45:00.995346 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fdd4eb5_f03c_4ba8_bf38_ca1a436f2eb5.slice/crio-a687c2d6589605bb94fff8e7d9b37f63bb1c395369d6a768be2638f2399a3d8e WatchSource:0}: Error finding container a687c2d6589605bb94fff8e7d9b37f63bb1c395369d6a768be2638f2399a3d8e: Status 404 returned error can't find the container with id a687c2d6589605bb94fff8e7d9b37f63bb1c395369d6a768be2638f2399a3d8e Dec 03 17:45:00 crc kubenswrapper[4998]: I1203 17:45:00.996206 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413065-8rf5w"] Dec 03 17:45:01 crc kubenswrapper[4998]: I1203 17:45:01.521793 4998 generic.go:334] "Generic (PLEG): container finished" podID="5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5" containerID="c6bb861fd30a36d1c3ae2a89cb480bf7f3b0ec8ce85faa9dcefbda762e0885e6" exitCode=0 Dec 03 17:45:01 crc kubenswrapper[4998]: I1203 17:45:01.521980 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8rf5w" event={"ID":"5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5","Type":"ContainerDied","Data":"c6bb861fd30a36d1c3ae2a89cb480bf7f3b0ec8ce85faa9dcefbda762e0885e6"} Dec 03 17:45:01 crc kubenswrapper[4998]: I1203 17:45:01.522087 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8rf5w" event={"ID":"5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5","Type":"ContainerStarted","Data":"a687c2d6589605bb94fff8e7d9b37f63bb1c395369d6a768be2638f2399a3d8e"} Dec 03 17:45:02 crc kubenswrapper[4998]: I1203 17:45:02.898289 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8rf5w" Dec 03 17:45:03 crc kubenswrapper[4998]: I1203 17:45:03.035434 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5-config-volume\") pod \"5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5\" (UID: \"5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5\") " Dec 03 17:45:03 crc kubenswrapper[4998]: I1203 17:45:03.036359 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5-secret-volume\") pod \"5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5\" (UID: \"5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5\") " Dec 03 17:45:03 crc kubenswrapper[4998]: I1203 17:45:03.036284 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5-config-volume" (OuterVolumeSpecName: "config-volume") pod "5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5" (UID: "5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:45:03 crc kubenswrapper[4998]: I1203 17:45:03.036590 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwffz\" (UniqueName: \"kubernetes.io/projected/5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5-kube-api-access-hwffz\") pod \"5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5\" (UID: \"5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5\") " Dec 03 17:45:03 crc kubenswrapper[4998]: I1203 17:45:03.038177 4998 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 17:45:03 crc kubenswrapper[4998]: I1203 17:45:03.059989 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5" (UID: "5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:45:03 crc kubenswrapper[4998]: I1203 17:45:03.060046 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5-kube-api-access-hwffz" (OuterVolumeSpecName: "kube-api-access-hwffz") pod "5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5" (UID: "5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5"). InnerVolumeSpecName "kube-api-access-hwffz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:45:03 crc kubenswrapper[4998]: I1203 17:45:03.162330 4998 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 17:45:03 crc kubenswrapper[4998]: I1203 17:45:03.162362 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwffz\" (UniqueName: \"kubernetes.io/projected/5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5-kube-api-access-hwffz\") on node \"crc\" DevicePath \"\"" Dec 03 17:45:03 crc kubenswrapper[4998]: I1203 17:45:03.543154 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8rf5w" event={"ID":"5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5","Type":"ContainerDied","Data":"a687c2d6589605bb94fff8e7d9b37f63bb1c395369d6a768be2638f2399a3d8e"} Dec 03 17:45:03 crc kubenswrapper[4998]: I1203 17:45:03.543702 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a687c2d6589605bb94fff8e7d9b37f63bb1c395369d6a768be2638f2399a3d8e" Dec 03 17:45:03 crc kubenswrapper[4998]: I1203 17:45:03.543239 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8rf5w" Dec 03 17:45:03 crc kubenswrapper[4998]: I1203 17:45:03.981386 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413020-j87g2"] Dec 03 17:45:03 crc kubenswrapper[4998]: I1203 17:45:03.991657 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413020-j87g2"] Dec 03 17:45:05 crc kubenswrapper[4998]: I1203 17:45:05.692852 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="026f9ce6-dbab-4142-8c66-3035f53b8054" path="/var/lib/kubelet/pods/026f9ce6-dbab-4142-8c66-3035f53b8054/volumes" Dec 03 17:45:27 crc kubenswrapper[4998]: I1203 17:45:27.110804 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:45:27 crc kubenswrapper[4998]: I1203 17:45:27.111354 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:45:27 crc kubenswrapper[4998]: I1203 17:45:27.111402 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 17:45:27 crc kubenswrapper[4998]: I1203 17:45:27.112357 4998 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c32ad81ca8fbe8808005dc071a77b0419d9a93bdb054bce341dafb8dbf1dd822"} pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:45:27 crc kubenswrapper[4998]: I1203 17:45:27.112442 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" containerID="cri-o://c32ad81ca8fbe8808005dc071a77b0419d9a93bdb054bce341dafb8dbf1dd822" gracePeriod=600 Dec 03 17:45:27 crc kubenswrapper[4998]: E1203 17:45:27.233171 4998 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27aeb142_d4e3_4827_ac1f_c34de6822b14.slice/crio-conmon-c32ad81ca8fbe8808005dc071a77b0419d9a93bdb054bce341dafb8dbf1dd822.scope\": RecentStats: unable to find data in memory cache]" Dec 03 17:45:27 crc kubenswrapper[4998]: I1203 17:45:27.793531 4998 generic.go:334] "Generic (PLEG): container finished" podID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerID="c32ad81ca8fbe8808005dc071a77b0419d9a93bdb054bce341dafb8dbf1dd822" exitCode=0 Dec 03 17:45:27 crc kubenswrapper[4998]: I1203 17:45:27.793719 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerDied","Data":"c32ad81ca8fbe8808005dc071a77b0419d9a93bdb054bce341dafb8dbf1dd822"} Dec 03 17:45:27 crc kubenswrapper[4998]: I1203 17:45:27.794050 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315"} Dec 03 17:45:27 crc kubenswrapper[4998]: I1203 17:45:27.794086 4998 scope.go:117] "RemoveContainer" containerID="7a9ebee7df6cb3ad8b72699c02ca0c2ec0b9fbba81f126c3e376b217237868a4" Dec 03 17:45:31 crc kubenswrapper[4998]: I1203 17:45:31.840040 4998 generic.go:334] "Generic (PLEG): container finished" podID="df3bf50a-6c2f-421e-8e35-3bcc2d368c6e" containerID="043c895a485290d169a5f9c1316bc03c62bc355e855ead7fed014c4a3b5e296a" exitCode=0 Dec 03 17:45:31 crc kubenswrapper[4998]: I1203 17:45:31.840116 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jx2sk/crc-debug-2rj8m" event={"ID":"df3bf50a-6c2f-421e-8e35-3bcc2d368c6e","Type":"ContainerDied","Data":"043c895a485290d169a5f9c1316bc03c62bc355e855ead7fed014c4a3b5e296a"} Dec 03 17:45:32 crc kubenswrapper[4998]: I1203 17:45:32.972203 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jx2sk/crc-debug-2rj8m" Dec 03 17:45:33 crc kubenswrapper[4998]: I1203 17:45:33.006472 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jx2sk/crc-debug-2rj8m"] Dec 03 17:45:33 crc kubenswrapper[4998]: I1203 17:45:33.015907 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jx2sk/crc-debug-2rj8m"] Dec 03 17:45:33 crc kubenswrapper[4998]: I1203 17:45:33.129686 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/df3bf50a-6c2f-421e-8e35-3bcc2d368c6e-host\") pod \"df3bf50a-6c2f-421e-8e35-3bcc2d368c6e\" (UID: \"df3bf50a-6c2f-421e-8e35-3bcc2d368c6e\") " Dec 03 17:45:33 crc kubenswrapper[4998]: I1203 17:45:33.129788 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6t95\" (UniqueName: \"kubernetes.io/projected/df3bf50a-6c2f-421e-8e35-3bcc2d368c6e-kube-api-access-k6t95\") pod \"df3bf50a-6c2f-421e-8e35-3bcc2d368c6e\" (UID: \"df3bf50a-6c2f-421e-8e35-3bcc2d368c6e\") " Dec 03 17:45:33 crc kubenswrapper[4998]: I1203 17:45:33.129957 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df3bf50a-6c2f-421e-8e35-3bcc2d368c6e-host" (OuterVolumeSpecName: "host") pod "df3bf50a-6c2f-421e-8e35-3bcc2d368c6e" (UID: "df3bf50a-6c2f-421e-8e35-3bcc2d368c6e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:45:33 crc kubenswrapper[4998]: I1203 17:45:33.131575 4998 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/df3bf50a-6c2f-421e-8e35-3bcc2d368c6e-host\") on node \"crc\" DevicePath \"\"" Dec 03 17:45:33 crc kubenswrapper[4998]: I1203 17:45:33.136395 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df3bf50a-6c2f-421e-8e35-3bcc2d368c6e-kube-api-access-k6t95" (OuterVolumeSpecName: "kube-api-access-k6t95") pod "df3bf50a-6c2f-421e-8e35-3bcc2d368c6e" (UID: "df3bf50a-6c2f-421e-8e35-3bcc2d368c6e"). InnerVolumeSpecName "kube-api-access-k6t95". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:45:33 crc kubenswrapper[4998]: I1203 17:45:33.234081 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6t95\" (UniqueName: \"kubernetes.io/projected/df3bf50a-6c2f-421e-8e35-3bcc2d368c6e-kube-api-access-k6t95\") on node \"crc\" DevicePath \"\"" Dec 03 17:45:33 crc kubenswrapper[4998]: I1203 17:45:33.689156 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df3bf50a-6c2f-421e-8e35-3bcc2d368c6e" path="/var/lib/kubelet/pods/df3bf50a-6c2f-421e-8e35-3bcc2d368c6e/volumes" Dec 03 17:45:33 crc kubenswrapper[4998]: I1203 17:45:33.864974 4998 scope.go:117] "RemoveContainer" containerID="043c895a485290d169a5f9c1316bc03c62bc355e855ead7fed014c4a3b5e296a" Dec 03 17:45:33 crc kubenswrapper[4998]: I1203 17:45:33.865042 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jx2sk/crc-debug-2rj8m" Dec 03 17:45:34 crc kubenswrapper[4998]: I1203 17:45:34.211677 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jx2sk/crc-debug-wdg7f"] Dec 03 17:45:34 crc kubenswrapper[4998]: E1203 17:45:34.212227 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5" containerName="collect-profiles" Dec 03 17:45:34 crc kubenswrapper[4998]: I1203 17:45:34.212242 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5" containerName="collect-profiles" Dec 03 17:45:34 crc kubenswrapper[4998]: E1203 17:45:34.212263 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df3bf50a-6c2f-421e-8e35-3bcc2d368c6e" containerName="container-00" Dec 03 17:45:34 crc kubenswrapper[4998]: I1203 17:45:34.212272 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="df3bf50a-6c2f-421e-8e35-3bcc2d368c6e" containerName="container-00" Dec 03 17:45:34 crc kubenswrapper[4998]: I1203 17:45:34.212615 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="df3bf50a-6c2f-421e-8e35-3bcc2d368c6e" containerName="container-00" Dec 03 17:45:34 crc kubenswrapper[4998]: I1203 17:45:34.212648 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fdd4eb5-f03c-4ba8-bf38-ca1a436f2eb5" containerName="collect-profiles" Dec 03 17:45:34 crc kubenswrapper[4998]: I1203 17:45:34.213620 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jx2sk/crc-debug-wdg7f" Dec 03 17:45:34 crc kubenswrapper[4998]: I1203 17:45:34.271337 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/713c9195-9105-434d-aed8-7267993e1780-host\") pod \"crc-debug-wdg7f\" (UID: \"713c9195-9105-434d-aed8-7267993e1780\") " pod="openshift-must-gather-jx2sk/crc-debug-wdg7f" Dec 03 17:45:34 crc kubenswrapper[4998]: I1203 17:45:34.271638 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqgb8\" (UniqueName: \"kubernetes.io/projected/713c9195-9105-434d-aed8-7267993e1780-kube-api-access-tqgb8\") pod \"crc-debug-wdg7f\" (UID: \"713c9195-9105-434d-aed8-7267993e1780\") " pod="openshift-must-gather-jx2sk/crc-debug-wdg7f" Dec 03 17:45:34 crc kubenswrapper[4998]: I1203 17:45:34.373545 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/713c9195-9105-434d-aed8-7267993e1780-host\") pod \"crc-debug-wdg7f\" (UID: \"713c9195-9105-434d-aed8-7267993e1780\") " pod="openshift-must-gather-jx2sk/crc-debug-wdg7f" Dec 03 17:45:34 crc kubenswrapper[4998]: I1203 17:45:34.373654 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqgb8\" (UniqueName: \"kubernetes.io/projected/713c9195-9105-434d-aed8-7267993e1780-kube-api-access-tqgb8\") pod \"crc-debug-wdg7f\" (UID: \"713c9195-9105-434d-aed8-7267993e1780\") " pod="openshift-must-gather-jx2sk/crc-debug-wdg7f" Dec 03 17:45:34 crc kubenswrapper[4998]: I1203 17:45:34.373945 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/713c9195-9105-434d-aed8-7267993e1780-host\") pod \"crc-debug-wdg7f\" (UID: \"713c9195-9105-434d-aed8-7267993e1780\") " pod="openshift-must-gather-jx2sk/crc-debug-wdg7f" Dec 03 17:45:34 crc kubenswrapper[4998]: I1203 17:45:34.394421 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqgb8\" (UniqueName: \"kubernetes.io/projected/713c9195-9105-434d-aed8-7267993e1780-kube-api-access-tqgb8\") pod \"crc-debug-wdg7f\" (UID: \"713c9195-9105-434d-aed8-7267993e1780\") " pod="openshift-must-gather-jx2sk/crc-debug-wdg7f" Dec 03 17:45:34 crc kubenswrapper[4998]: I1203 17:45:34.528717 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jx2sk/crc-debug-wdg7f" Dec 03 17:45:34 crc kubenswrapper[4998]: I1203 17:45:34.880327 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jx2sk/crc-debug-wdg7f" event={"ID":"713c9195-9105-434d-aed8-7267993e1780","Type":"ContainerStarted","Data":"edc92c43074e48fd4d5e91975d1114819bdfced6b7eda4e5c4b10cb4d26d7547"} Dec 03 17:45:35 crc kubenswrapper[4998]: I1203 17:45:35.890418 4998 generic.go:334] "Generic (PLEG): container finished" podID="713c9195-9105-434d-aed8-7267993e1780" containerID="c2c1eb09b44eaa1c8b68aeab5f3c6629419e507a86be42dd66ce687350f49d9d" exitCode=0 Dec 03 17:45:35 crc kubenswrapper[4998]: I1203 17:45:35.890487 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jx2sk/crc-debug-wdg7f" event={"ID":"713c9195-9105-434d-aed8-7267993e1780","Type":"ContainerDied","Data":"c2c1eb09b44eaa1c8b68aeab5f3c6629419e507a86be42dd66ce687350f49d9d"} Dec 03 17:45:37 crc kubenswrapper[4998]: I1203 17:45:37.020641 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jx2sk/crc-debug-wdg7f" Dec 03 17:45:37 crc kubenswrapper[4998]: I1203 17:45:37.125250 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/713c9195-9105-434d-aed8-7267993e1780-host\") pod \"713c9195-9105-434d-aed8-7267993e1780\" (UID: \"713c9195-9105-434d-aed8-7267993e1780\") " Dec 03 17:45:37 crc kubenswrapper[4998]: I1203 17:45:37.125307 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqgb8\" (UniqueName: \"kubernetes.io/projected/713c9195-9105-434d-aed8-7267993e1780-kube-api-access-tqgb8\") pod \"713c9195-9105-434d-aed8-7267993e1780\" (UID: \"713c9195-9105-434d-aed8-7267993e1780\") " Dec 03 17:45:37 crc kubenswrapper[4998]: I1203 17:45:37.125427 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/713c9195-9105-434d-aed8-7267993e1780-host" (OuterVolumeSpecName: "host") pod "713c9195-9105-434d-aed8-7267993e1780" (UID: "713c9195-9105-434d-aed8-7267993e1780"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:45:37 crc kubenswrapper[4998]: I1203 17:45:37.126814 4998 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/713c9195-9105-434d-aed8-7267993e1780-host\") on node \"crc\" DevicePath \"\"" Dec 03 17:45:37 crc kubenswrapper[4998]: I1203 17:45:37.147010 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/713c9195-9105-434d-aed8-7267993e1780-kube-api-access-tqgb8" (OuterVolumeSpecName: "kube-api-access-tqgb8") pod "713c9195-9105-434d-aed8-7267993e1780" (UID: "713c9195-9105-434d-aed8-7267993e1780"). InnerVolumeSpecName "kube-api-access-tqgb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:45:37 crc kubenswrapper[4998]: I1203 17:45:37.227740 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqgb8\" (UniqueName: \"kubernetes.io/projected/713c9195-9105-434d-aed8-7267993e1780-kube-api-access-tqgb8\") on node \"crc\" DevicePath \"\"" Dec 03 17:45:37 crc kubenswrapper[4998]: I1203 17:45:37.907816 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jx2sk/crc-debug-wdg7f" event={"ID":"713c9195-9105-434d-aed8-7267993e1780","Type":"ContainerDied","Data":"edc92c43074e48fd4d5e91975d1114819bdfced6b7eda4e5c4b10cb4d26d7547"} Dec 03 17:45:37 crc kubenswrapper[4998]: I1203 17:45:37.908080 4998 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="edc92c43074e48fd4d5e91975d1114819bdfced6b7eda4e5c4b10cb4d26d7547" Dec 03 17:45:37 crc kubenswrapper[4998]: I1203 17:45:37.908131 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jx2sk/crc-debug-wdg7f" Dec 03 17:45:38 crc kubenswrapper[4998]: I1203 17:45:38.274899 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jx2sk/crc-debug-wdg7f"] Dec 03 17:45:38 crc kubenswrapper[4998]: I1203 17:45:38.283372 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jx2sk/crc-debug-wdg7f"] Dec 03 17:45:39 crc kubenswrapper[4998]: I1203 17:45:39.526115 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jx2sk/crc-debug-v8bdm"] Dec 03 17:45:39 crc kubenswrapper[4998]: E1203 17:45:39.526882 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="713c9195-9105-434d-aed8-7267993e1780" containerName="container-00" Dec 03 17:45:39 crc kubenswrapper[4998]: I1203 17:45:39.526898 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="713c9195-9105-434d-aed8-7267993e1780" containerName="container-00" Dec 03 17:45:39 crc kubenswrapper[4998]: I1203 17:45:39.527106 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="713c9195-9105-434d-aed8-7267993e1780" containerName="container-00" Dec 03 17:45:39 crc kubenswrapper[4998]: I1203 17:45:39.527805 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jx2sk/crc-debug-v8bdm" Dec 03 17:45:39 crc kubenswrapper[4998]: I1203 17:45:39.573519 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r8ss\" (UniqueName: \"kubernetes.io/projected/40e93665-5393-4128-be89-d815e98685e0-kube-api-access-5r8ss\") pod \"crc-debug-v8bdm\" (UID: \"40e93665-5393-4128-be89-d815e98685e0\") " pod="openshift-must-gather-jx2sk/crc-debug-v8bdm" Dec 03 17:45:39 crc kubenswrapper[4998]: I1203 17:45:39.573656 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/40e93665-5393-4128-be89-d815e98685e0-host\") pod \"crc-debug-v8bdm\" (UID: \"40e93665-5393-4128-be89-d815e98685e0\") " pod="openshift-must-gather-jx2sk/crc-debug-v8bdm" Dec 03 17:45:39 crc kubenswrapper[4998]: I1203 17:45:39.676277 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r8ss\" (UniqueName: \"kubernetes.io/projected/40e93665-5393-4128-be89-d815e98685e0-kube-api-access-5r8ss\") pod \"crc-debug-v8bdm\" (UID: \"40e93665-5393-4128-be89-d815e98685e0\") " pod="openshift-must-gather-jx2sk/crc-debug-v8bdm" Dec 03 17:45:39 crc kubenswrapper[4998]: I1203 17:45:39.676388 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/40e93665-5393-4128-be89-d815e98685e0-host\") pod \"crc-debug-v8bdm\" (UID: \"40e93665-5393-4128-be89-d815e98685e0\") " pod="openshift-must-gather-jx2sk/crc-debug-v8bdm" Dec 03 17:45:39 crc kubenswrapper[4998]: I1203 17:45:39.676642 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/40e93665-5393-4128-be89-d815e98685e0-host\") pod \"crc-debug-v8bdm\" (UID: \"40e93665-5393-4128-be89-d815e98685e0\") " pod="openshift-must-gather-jx2sk/crc-debug-v8bdm" Dec 03 17:45:39 crc kubenswrapper[4998]: I1203 17:45:39.694885 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="713c9195-9105-434d-aed8-7267993e1780" path="/var/lib/kubelet/pods/713c9195-9105-434d-aed8-7267993e1780/volumes" Dec 03 17:45:39 crc kubenswrapper[4998]: I1203 17:45:39.704235 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r8ss\" (UniqueName: \"kubernetes.io/projected/40e93665-5393-4128-be89-d815e98685e0-kube-api-access-5r8ss\") pod \"crc-debug-v8bdm\" (UID: \"40e93665-5393-4128-be89-d815e98685e0\") " pod="openshift-must-gather-jx2sk/crc-debug-v8bdm" Dec 03 17:45:39 crc kubenswrapper[4998]: I1203 17:45:39.848880 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jx2sk/crc-debug-v8bdm" Dec 03 17:45:39 crc kubenswrapper[4998]: I1203 17:45:39.930387 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jx2sk/crc-debug-v8bdm" event={"ID":"40e93665-5393-4128-be89-d815e98685e0","Type":"ContainerStarted","Data":"b27d97150b62725a8185d15b890ab07bbaae4f1b9da0ae4227229ec2da3bbab0"} Dec 03 17:45:40 crc kubenswrapper[4998]: I1203 17:45:40.943849 4998 generic.go:334] "Generic (PLEG): container finished" podID="40e93665-5393-4128-be89-d815e98685e0" containerID="36eed79b90a2a189d032acadb3b1ce1b99011c855e342c803d556bff2e959f28" exitCode=0 Dec 03 17:45:40 crc kubenswrapper[4998]: I1203 17:45:40.944102 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jx2sk/crc-debug-v8bdm" event={"ID":"40e93665-5393-4128-be89-d815e98685e0","Type":"ContainerDied","Data":"36eed79b90a2a189d032acadb3b1ce1b99011c855e342c803d556bff2e959f28"} Dec 03 17:45:40 crc kubenswrapper[4998]: I1203 17:45:40.992449 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jx2sk/crc-debug-v8bdm"] Dec 03 17:45:41 crc kubenswrapper[4998]: I1203 17:45:41.001149 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jx2sk/crc-debug-v8bdm"] Dec 03 17:45:42 crc kubenswrapper[4998]: I1203 17:45:42.069880 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jx2sk/crc-debug-v8bdm" Dec 03 17:45:42 crc kubenswrapper[4998]: I1203 17:45:42.240615 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/40e93665-5393-4128-be89-d815e98685e0-host\") pod \"40e93665-5393-4128-be89-d815e98685e0\" (UID: \"40e93665-5393-4128-be89-d815e98685e0\") " Dec 03 17:45:42 crc kubenswrapper[4998]: I1203 17:45:42.240789 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5r8ss\" (UniqueName: \"kubernetes.io/projected/40e93665-5393-4128-be89-d815e98685e0-kube-api-access-5r8ss\") pod \"40e93665-5393-4128-be89-d815e98685e0\" (UID: \"40e93665-5393-4128-be89-d815e98685e0\") " Dec 03 17:45:42 crc kubenswrapper[4998]: I1203 17:45:42.240775 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40e93665-5393-4128-be89-d815e98685e0-host" (OuterVolumeSpecName: "host") pod "40e93665-5393-4128-be89-d815e98685e0" (UID: "40e93665-5393-4128-be89-d815e98685e0"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:45:42 crc kubenswrapper[4998]: I1203 17:45:42.241501 4998 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/40e93665-5393-4128-be89-d815e98685e0-host\") on node \"crc\" DevicePath \"\"" Dec 03 17:45:42 crc kubenswrapper[4998]: I1203 17:45:42.246937 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40e93665-5393-4128-be89-d815e98685e0-kube-api-access-5r8ss" (OuterVolumeSpecName: "kube-api-access-5r8ss") pod "40e93665-5393-4128-be89-d815e98685e0" (UID: "40e93665-5393-4128-be89-d815e98685e0"). InnerVolumeSpecName "kube-api-access-5r8ss". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:45:42 crc kubenswrapper[4998]: I1203 17:45:42.343077 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5r8ss\" (UniqueName: \"kubernetes.io/projected/40e93665-5393-4128-be89-d815e98685e0-kube-api-access-5r8ss\") on node \"crc\" DevicePath \"\"" Dec 03 17:45:42 crc kubenswrapper[4998]: I1203 17:45:42.969311 4998 scope.go:117] "RemoveContainer" containerID="36eed79b90a2a189d032acadb3b1ce1b99011c855e342c803d556bff2e959f28" Dec 03 17:45:42 crc kubenswrapper[4998]: I1203 17:45:42.969502 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jx2sk/crc-debug-v8bdm" Dec 03 17:45:43 crc kubenswrapper[4998]: I1203 17:45:43.688507 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40e93665-5393-4128-be89-d815e98685e0" path="/var/lib/kubelet/pods/40e93665-5393-4128-be89-d815e98685e0/volumes" Dec 03 17:45:46 crc kubenswrapper[4998]: I1203 17:45:46.081563 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-znppm"] Dec 03 17:45:46 crc kubenswrapper[4998]: E1203 17:45:46.083908 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40e93665-5393-4128-be89-d815e98685e0" containerName="container-00" Dec 03 17:45:46 crc kubenswrapper[4998]: I1203 17:45:46.083936 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="40e93665-5393-4128-be89-d815e98685e0" containerName="container-00" Dec 03 17:45:46 crc kubenswrapper[4998]: I1203 17:45:46.084199 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="40e93665-5393-4128-be89-d815e98685e0" containerName="container-00" Dec 03 17:45:46 crc kubenswrapper[4998]: I1203 17:45:46.085841 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-znppm" Dec 03 17:45:46 crc kubenswrapper[4998]: I1203 17:45:46.094418 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-znppm"] Dec 03 17:45:46 crc kubenswrapper[4998]: I1203 17:45:46.237271 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad81770d-9852-4abb-97b6-c803bce55972-catalog-content\") pod \"redhat-operators-znppm\" (UID: \"ad81770d-9852-4abb-97b6-c803bce55972\") " pod="openshift-marketplace/redhat-operators-znppm" Dec 03 17:45:46 crc kubenswrapper[4998]: I1203 17:45:46.238550 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhnfb\" (UniqueName: \"kubernetes.io/projected/ad81770d-9852-4abb-97b6-c803bce55972-kube-api-access-rhnfb\") pod \"redhat-operators-znppm\" (UID: \"ad81770d-9852-4abb-97b6-c803bce55972\") " pod="openshift-marketplace/redhat-operators-znppm" Dec 03 17:45:46 crc kubenswrapper[4998]: I1203 17:45:46.238707 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad81770d-9852-4abb-97b6-c803bce55972-utilities\") pod \"redhat-operators-znppm\" (UID: \"ad81770d-9852-4abb-97b6-c803bce55972\") " pod="openshift-marketplace/redhat-operators-znppm" Dec 03 17:45:46 crc kubenswrapper[4998]: I1203 17:45:46.340179 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad81770d-9852-4abb-97b6-c803bce55972-catalog-content\") pod \"redhat-operators-znppm\" (UID: \"ad81770d-9852-4abb-97b6-c803bce55972\") " pod="openshift-marketplace/redhat-operators-znppm" Dec 03 17:45:46 crc kubenswrapper[4998]: I1203 17:45:46.340260 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhnfb\" (UniqueName: \"kubernetes.io/projected/ad81770d-9852-4abb-97b6-c803bce55972-kube-api-access-rhnfb\") pod \"redhat-operators-znppm\" (UID: \"ad81770d-9852-4abb-97b6-c803bce55972\") " pod="openshift-marketplace/redhat-operators-znppm" Dec 03 17:45:46 crc kubenswrapper[4998]: I1203 17:45:46.340307 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad81770d-9852-4abb-97b6-c803bce55972-utilities\") pod \"redhat-operators-znppm\" (UID: \"ad81770d-9852-4abb-97b6-c803bce55972\") " pod="openshift-marketplace/redhat-operators-znppm" Dec 03 17:45:46 crc kubenswrapper[4998]: I1203 17:45:46.340685 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad81770d-9852-4abb-97b6-c803bce55972-catalog-content\") pod \"redhat-operators-znppm\" (UID: \"ad81770d-9852-4abb-97b6-c803bce55972\") " pod="openshift-marketplace/redhat-operators-znppm" Dec 03 17:45:46 crc kubenswrapper[4998]: I1203 17:45:46.340725 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad81770d-9852-4abb-97b6-c803bce55972-utilities\") pod \"redhat-operators-znppm\" (UID: \"ad81770d-9852-4abb-97b6-c803bce55972\") " pod="openshift-marketplace/redhat-operators-znppm" Dec 03 17:45:46 crc kubenswrapper[4998]: I1203 17:45:46.365082 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhnfb\" (UniqueName: \"kubernetes.io/projected/ad81770d-9852-4abb-97b6-c803bce55972-kube-api-access-rhnfb\") pod \"redhat-operators-znppm\" (UID: \"ad81770d-9852-4abb-97b6-c803bce55972\") " pod="openshift-marketplace/redhat-operators-znppm" Dec 03 17:45:46 crc kubenswrapper[4998]: I1203 17:45:46.403588 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-znppm" Dec 03 17:45:46 crc kubenswrapper[4998]: I1203 17:45:46.891701 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-znppm"] Dec 03 17:45:46 crc kubenswrapper[4998]: W1203 17:45:46.899211 4998 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad81770d_9852_4abb_97b6_c803bce55972.slice/crio-4092d064f9965ff42393b0b4205c72d4032620318eaba551e24f34f8d815cdbb WatchSource:0}: Error finding container 4092d064f9965ff42393b0b4205c72d4032620318eaba551e24f34f8d815cdbb: Status 404 returned error can't find the container with id 4092d064f9965ff42393b0b4205c72d4032620318eaba551e24f34f8d815cdbb Dec 03 17:45:47 crc kubenswrapper[4998]: I1203 17:45:47.007860 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-znppm" event={"ID":"ad81770d-9852-4abb-97b6-c803bce55972","Type":"ContainerStarted","Data":"4092d064f9965ff42393b0b4205c72d4032620318eaba551e24f34f8d815cdbb"} Dec 03 17:45:48 crc kubenswrapper[4998]: I1203 17:45:48.022554 4998 generic.go:334] "Generic (PLEG): container finished" podID="ad81770d-9852-4abb-97b6-c803bce55972" containerID="8a8222bc2b37c2015049b2b42d280f18791b62eece913ae6234ceb9a7d1a67d7" exitCode=0 Dec 03 17:45:48 crc kubenswrapper[4998]: I1203 17:45:48.022906 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-znppm" event={"ID":"ad81770d-9852-4abb-97b6-c803bce55972","Type":"ContainerDied","Data":"8a8222bc2b37c2015049b2b42d280f18791b62eece913ae6234ceb9a7d1a67d7"} Dec 03 17:45:48 crc kubenswrapper[4998]: I1203 17:45:48.024911 4998 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 17:45:49 crc kubenswrapper[4998]: I1203 17:45:49.034306 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-znppm" event={"ID":"ad81770d-9852-4abb-97b6-c803bce55972","Type":"ContainerStarted","Data":"b5b77fe977aca2443978238630c9beef595d6b5274dbfd0599ca8b3f0ae9ed95"} Dec 03 17:45:53 crc kubenswrapper[4998]: I1203 17:45:53.075644 4998 generic.go:334] "Generic (PLEG): container finished" podID="ad81770d-9852-4abb-97b6-c803bce55972" containerID="b5b77fe977aca2443978238630c9beef595d6b5274dbfd0599ca8b3f0ae9ed95" exitCode=0 Dec 03 17:45:53 crc kubenswrapper[4998]: I1203 17:45:53.076003 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-znppm" event={"ID":"ad81770d-9852-4abb-97b6-c803bce55972","Type":"ContainerDied","Data":"b5b77fe977aca2443978238630c9beef595d6b5274dbfd0599ca8b3f0ae9ed95"} Dec 03 17:45:55 crc kubenswrapper[4998]: I1203 17:45:55.100081 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-znppm" event={"ID":"ad81770d-9852-4abb-97b6-c803bce55972","Type":"ContainerStarted","Data":"bda8f8d82a43cd414c1a347f26c5f99e3919287487e06ea2a2f082f48115d5d9"} Dec 03 17:45:55 crc kubenswrapper[4998]: I1203 17:45:55.125750 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-znppm" podStartSLOduration=2.762233043 podStartE2EDuration="9.125733282s" podCreationTimestamp="2025-12-03 17:45:46 +0000 UTC" firstStartedPulling="2025-12-03 17:45:48.024675288 +0000 UTC m=+6126.636375511" lastFinishedPulling="2025-12-03 17:45:54.388175517 +0000 UTC m=+6132.999875750" observedRunningTime="2025-12-03 17:45:55.117125598 +0000 UTC m=+6133.728825821" watchObservedRunningTime="2025-12-03 17:45:55.125733282 +0000 UTC m=+6133.737433505" Dec 03 17:45:56 crc kubenswrapper[4998]: I1203 17:45:56.404013 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-znppm" Dec 03 17:45:56 crc kubenswrapper[4998]: I1203 17:45:56.404449 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-znppm" Dec 03 17:45:57 crc kubenswrapper[4998]: I1203 17:45:57.466976 4998 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-znppm" podUID="ad81770d-9852-4abb-97b6-c803bce55972" containerName="registry-server" probeResult="failure" output=< Dec 03 17:45:57 crc kubenswrapper[4998]: timeout: failed to connect service ":50051" within 1s Dec 03 17:45:57 crc kubenswrapper[4998]: > Dec 03 17:46:03 crc kubenswrapper[4998]: I1203 17:46:03.252111 4998 scope.go:117] "RemoveContainer" containerID="80e43abb86da11c9f5c02522e86ce07e1497f34e8a79d05fbd7e93e9c2997bbf" Dec 03 17:46:06 crc kubenswrapper[4998]: I1203 17:46:06.464413 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-znppm" Dec 03 17:46:06 crc kubenswrapper[4998]: I1203 17:46:06.523303 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-znppm" Dec 03 17:46:07 crc kubenswrapper[4998]: I1203 17:46:07.271897 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-57f6856fd8-t44p5_e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4/barbican-api/0.log" Dec 03 17:46:07 crc kubenswrapper[4998]: I1203 17:46:07.314634 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-57f6856fd8-t44p5_e5a7538a-c4f6-45ba-8325-b75a5ecc1fb4/barbican-api-log/0.log" Dec 03 17:46:07 crc kubenswrapper[4998]: I1203 17:46:07.353591 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-znppm"] Dec 03 17:46:07 crc kubenswrapper[4998]: I1203 17:46:07.437930 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7d588c5fc8-x85fx_57638b87-e1dd-4583-acc1-d57eb123de4d/barbican-keystone-listener/0.log" Dec 03 17:46:07 crc kubenswrapper[4998]: I1203 17:46:07.555454 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7d588c5fc8-x85fx_57638b87-e1dd-4583-acc1-d57eb123de4d/barbican-keystone-listener-log/0.log" Dec 03 17:46:07 crc kubenswrapper[4998]: I1203 17:46:07.624877 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7d6d5bf649-h8gnx_d33fe6f6-a02c-4841-baf4-1c956cd2e9f2/barbican-worker/0.log" Dec 03 17:46:07 crc kubenswrapper[4998]: I1203 17:46:07.710122 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7d6d5bf649-h8gnx_d33fe6f6-a02c-4841-baf4-1c956cd2e9f2/barbican-worker-log/0.log" Dec 03 17:46:07 crc kubenswrapper[4998]: I1203 17:46:07.863138 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-tz6w6_e700335e-e6da-4bd2-ad00-6a7ecb9761a0/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 17:46:08 crc kubenswrapper[4998]: I1203 17:46:08.199001 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_fa21daa5-d0b4-4d69-a959-f79a5869d9fe/proxy-httpd/0.log" Dec 03 17:46:08 crc kubenswrapper[4998]: I1203 17:46:08.205476 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_fa21daa5-d0b4-4d69-a959-f79a5869d9fe/ceilometer-notification-agent/0.log" Dec 03 17:46:08 crc kubenswrapper[4998]: I1203 17:46:08.238501 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_fa21daa5-d0b4-4d69-a959-f79a5869d9fe/ceilometer-central-agent/0.log" Dec 03 17:46:08 crc kubenswrapper[4998]: I1203 17:46:08.252180 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-znppm" podUID="ad81770d-9852-4abb-97b6-c803bce55972" containerName="registry-server" containerID="cri-o://bda8f8d82a43cd414c1a347f26c5f99e3919287487e06ea2a2f082f48115d5d9" gracePeriod=2 Dec 03 17:46:08 crc kubenswrapper[4998]: I1203 17:46:08.364921 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_fa21daa5-d0b4-4d69-a959-f79a5869d9fe/sg-core/0.log" Dec 03 17:46:08 crc kubenswrapper[4998]: I1203 17:46:08.545089 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_937de507-f2cb-4b8c-8b86-cfc929f23248/cinder-api-log/0.log" Dec 03 17:46:08 crc kubenswrapper[4998]: I1203 17:46:08.810948 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-znppm" Dec 03 17:46:08 crc kubenswrapper[4998]: I1203 17:46:08.907974 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_0f0ad14b-5286-4d27-8dd1-0a4df70120a6/probe/0.log" Dec 03 17:46:08 crc kubenswrapper[4998]: I1203 17:46:08.952088 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad81770d-9852-4abb-97b6-c803bce55972-catalog-content\") pod \"ad81770d-9852-4abb-97b6-c803bce55972\" (UID: \"ad81770d-9852-4abb-97b6-c803bce55972\") " Dec 03 17:46:08 crc kubenswrapper[4998]: I1203 17:46:08.952396 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhnfb\" (UniqueName: \"kubernetes.io/projected/ad81770d-9852-4abb-97b6-c803bce55972-kube-api-access-rhnfb\") pod \"ad81770d-9852-4abb-97b6-c803bce55972\" (UID: \"ad81770d-9852-4abb-97b6-c803bce55972\") " Dec 03 17:46:08 crc kubenswrapper[4998]: I1203 17:46:08.952546 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad81770d-9852-4abb-97b6-c803bce55972-utilities\") pod \"ad81770d-9852-4abb-97b6-c803bce55972\" (UID: \"ad81770d-9852-4abb-97b6-c803bce55972\") " Dec 03 17:46:08 crc kubenswrapper[4998]: I1203 17:46:08.953376 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad81770d-9852-4abb-97b6-c803bce55972-utilities" (OuterVolumeSpecName: "utilities") pod "ad81770d-9852-4abb-97b6-c803bce55972" (UID: "ad81770d-9852-4abb-97b6-c803bce55972"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:46:08 crc kubenswrapper[4998]: I1203 17:46:08.958569 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad81770d-9852-4abb-97b6-c803bce55972-kube-api-access-rhnfb" (OuterVolumeSpecName: "kube-api-access-rhnfb") pod "ad81770d-9852-4abb-97b6-c803bce55972" (UID: "ad81770d-9852-4abb-97b6-c803bce55972"). InnerVolumeSpecName "kube-api-access-rhnfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.054194 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_937de507-f2cb-4b8c-8b86-cfc929f23248/cinder-api/0.log" Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.054724 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad81770d-9852-4abb-97b6-c803bce55972-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.054792 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhnfb\" (UniqueName: \"kubernetes.io/projected/ad81770d-9852-4abb-97b6-c803bce55972-kube-api-access-rhnfb\") on node \"crc\" DevicePath \"\"" Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.065429 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad81770d-9852-4abb-97b6-c803bce55972-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad81770d-9852-4abb-97b6-c803bce55972" (UID: "ad81770d-9852-4abb-97b6-c803bce55972"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.104191 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_0f0ad14b-5286-4d27-8dd1-0a4df70120a6/cinder-backup/0.log" Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.146938 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_4642ed6c-0202-4a7d-b50d-9c90247cfdb9/cinder-scheduler/0.log" Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.156622 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad81770d-9852-4abb-97b6-c803bce55972-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.287163 4998 generic.go:334] "Generic (PLEG): container finished" podID="ad81770d-9852-4abb-97b6-c803bce55972" containerID="bda8f8d82a43cd414c1a347f26c5f99e3919287487e06ea2a2f082f48115d5d9" exitCode=0 Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.287218 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-znppm" event={"ID":"ad81770d-9852-4abb-97b6-c803bce55972","Type":"ContainerDied","Data":"bda8f8d82a43cd414c1a347f26c5f99e3919287487e06ea2a2f082f48115d5d9"} Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.287250 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-znppm" event={"ID":"ad81770d-9852-4abb-97b6-c803bce55972","Type":"ContainerDied","Data":"4092d064f9965ff42393b0b4205c72d4032620318eaba551e24f34f8d815cdbb"} Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.287272 4998 scope.go:117] "RemoveContainer" containerID="bda8f8d82a43cd414c1a347f26c5f99e3919287487e06ea2a2f082f48115d5d9" Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.287417 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-znppm" Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.288718 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_4642ed6c-0202-4a7d-b50d-9c90247cfdb9/probe/0.log" Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.307463 4998 scope.go:117] "RemoveContainer" containerID="b5b77fe977aca2443978238630c9beef595d6b5274dbfd0599ca8b3f0ae9ed95" Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.330849 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-znppm"] Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.335651 4998 scope.go:117] "RemoveContainer" containerID="8a8222bc2b37c2015049b2b42d280f18791b62eece913ae6234ceb9a7d1a67d7" Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.345436 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-znppm"] Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.386956 4998 scope.go:117] "RemoveContainer" containerID="bda8f8d82a43cd414c1a347f26c5f99e3919287487e06ea2a2f082f48115d5d9" Dec 03 17:46:09 crc kubenswrapper[4998]: E1203 17:46:09.387406 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bda8f8d82a43cd414c1a347f26c5f99e3919287487e06ea2a2f082f48115d5d9\": container with ID starting with bda8f8d82a43cd414c1a347f26c5f99e3919287487e06ea2a2f082f48115d5d9 not found: ID does not exist" containerID="bda8f8d82a43cd414c1a347f26c5f99e3919287487e06ea2a2f082f48115d5d9" Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.387503 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bda8f8d82a43cd414c1a347f26c5f99e3919287487e06ea2a2f082f48115d5d9"} err="failed to get container status \"bda8f8d82a43cd414c1a347f26c5f99e3919287487e06ea2a2f082f48115d5d9\": rpc error: code = NotFound desc = could not find container \"bda8f8d82a43cd414c1a347f26c5f99e3919287487e06ea2a2f082f48115d5d9\": container with ID starting with bda8f8d82a43cd414c1a347f26c5f99e3919287487e06ea2a2f082f48115d5d9 not found: ID does not exist" Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.387583 4998 scope.go:117] "RemoveContainer" containerID="b5b77fe977aca2443978238630c9beef595d6b5274dbfd0599ca8b3f0ae9ed95" Dec 03 17:46:09 crc kubenswrapper[4998]: E1203 17:46:09.388129 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5b77fe977aca2443978238630c9beef595d6b5274dbfd0599ca8b3f0ae9ed95\": container with ID starting with b5b77fe977aca2443978238630c9beef595d6b5274dbfd0599ca8b3f0ae9ed95 not found: ID does not exist" containerID="b5b77fe977aca2443978238630c9beef595d6b5274dbfd0599ca8b3f0ae9ed95" Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.388181 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5b77fe977aca2443978238630c9beef595d6b5274dbfd0599ca8b3f0ae9ed95"} err="failed to get container status \"b5b77fe977aca2443978238630c9beef595d6b5274dbfd0599ca8b3f0ae9ed95\": rpc error: code = NotFound desc = could not find container \"b5b77fe977aca2443978238630c9beef595d6b5274dbfd0599ca8b3f0ae9ed95\": container with ID starting with b5b77fe977aca2443978238630c9beef595d6b5274dbfd0599ca8b3f0ae9ed95 not found: ID does not exist" Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.388215 4998 scope.go:117] "RemoveContainer" containerID="8a8222bc2b37c2015049b2b42d280f18791b62eece913ae6234ceb9a7d1a67d7" Dec 03 17:46:09 crc kubenswrapper[4998]: E1203 17:46:09.388624 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a8222bc2b37c2015049b2b42d280f18791b62eece913ae6234ceb9a7d1a67d7\": container with ID starting with 8a8222bc2b37c2015049b2b42d280f18791b62eece913ae6234ceb9a7d1a67d7 not found: ID does not exist" containerID="8a8222bc2b37c2015049b2b42d280f18791b62eece913ae6234ceb9a7d1a67d7" Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.388711 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a8222bc2b37c2015049b2b42d280f18791b62eece913ae6234ceb9a7d1a67d7"} err="failed to get container status \"8a8222bc2b37c2015049b2b42d280f18791b62eece913ae6234ceb9a7d1a67d7\": rpc error: code = NotFound desc = could not find container \"8a8222bc2b37c2015049b2b42d280f18791b62eece913ae6234ceb9a7d1a67d7\": container with ID starting with 8a8222bc2b37c2015049b2b42d280f18791b62eece913ae6234ceb9a7d1a67d7 not found: ID does not exist" Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.415162 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-0_f9f63cb2-129d-428c-9a3b-f7646f3c8eb1/probe/0.log" Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.691380 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-0_f9f63cb2-129d-428c-9a3b-f7646f3c8eb1/cinder-volume/0.log" Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.703023 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad81770d-9852-4abb-97b6-c803bce55972" path="/var/lib/kubelet/pods/ad81770d-9852-4abb-97b6-c803bce55972/volumes" Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.777232 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-2-0_d83ce9f0-9df6-470e-8dc5-ea647fd082ca/probe/0.log" Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.798615 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-nfs-2-0_d83ce9f0-9df6-470e-8dc5-ea647fd082ca/cinder-volume/0.log" Dec 03 17:46:09 crc kubenswrapper[4998]: I1203 17:46:09.923257 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-dfqtk_5fb49d67-f5a0-4728-b3e1-dc596955f03a/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 17:46:10 crc kubenswrapper[4998]: I1203 17:46:10.013140 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-v57vv_4bd61b3c-44d3-46c3-9e4e-094937642ef8/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 17:46:10 crc kubenswrapper[4998]: I1203 17:46:10.160224 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-d7878c9d7-b9xcm_3b8aca52-64f7-41bc-a1a1-0ae5ef118606/init/0.log" Dec 03 17:46:10 crc kubenswrapper[4998]: I1203 17:46:10.311652 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-d7878c9d7-b9xcm_3b8aca52-64f7-41bc-a1a1-0ae5ef118606/init/0.log" Dec 03 17:46:10 crc kubenswrapper[4998]: I1203 17:46:10.407558 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-s2f8q_742311fa-8f91-4738-ae24-66e591d9b8a1/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 17:46:10 crc kubenswrapper[4998]: I1203 17:46:10.475812 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-d7878c9d7-b9xcm_3b8aca52-64f7-41bc-a1a1-0ae5ef118606/dnsmasq-dns/0.log" Dec 03 17:46:10 crc kubenswrapper[4998]: I1203 17:46:10.617516 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_1cbdcfac-d2fe-456c-9632-f67aca0d05d5/glance-httpd/0.log" Dec 03 17:46:10 crc kubenswrapper[4998]: I1203 17:46:10.648340 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_1cbdcfac-d2fe-456c-9632-f67aca0d05d5/glance-log/0.log" Dec 03 17:46:10 crc kubenswrapper[4998]: I1203 17:46:10.780381 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7874dd07-6286-4514-8896-309a007a9aee/glance-httpd/0.log" Dec 03 17:46:10 crc kubenswrapper[4998]: I1203 17:46:10.840300 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7874dd07-6286-4514-8896-309a007a9aee/glance-log/0.log" Dec 03 17:46:11 crc kubenswrapper[4998]: I1203 17:46:11.016088 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7c5f7f9446-f97x8_24f1f198-c11a-4b5d-bcdd-d359ed88d97f/horizon/0.log" Dec 03 17:46:11 crc kubenswrapper[4998]: I1203 17:46:11.120206 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-ffvcf_c5cebbf5-5350-42c2-8908-8a508995a72c/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 17:46:11 crc kubenswrapper[4998]: I1203 17:46:11.324041 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-g8wgf_6f58e5a8-bad5-4d0e-ae33-8e666718bf0a/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 17:46:11 crc kubenswrapper[4998]: I1203 17:46:11.602034 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29413021-8jrpv_526572d2-2bee-415c-a31c-def5dd4f520a/keystone-cron/0.log" Dec 03 17:46:11 crc kubenswrapper[4998]: I1203 17:46:11.698684 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_84df7dec-fa0b-4b79-8e59-5298650093e5/kube-state-metrics/0.log" Dec 03 17:46:11 crc kubenswrapper[4998]: I1203 17:46:11.830359 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7c5f7f9446-f97x8_24f1f198-c11a-4b5d-bcdd-d359ed88d97f/horizon-log/0.log" Dec 03 17:46:11 crc kubenswrapper[4998]: I1203 17:46:11.937208 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-s6x8q_d23ecf9c-0347-4a37-ae82-c7b24d5b9575/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 17:46:11 crc kubenswrapper[4998]: I1203 17:46:11.973614 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6764c5c65f-jtlz7_d245e0d1-7808-4614-870e-64947ae9fb77/keystone-api/0.log" Dec 03 17:46:12 crc kubenswrapper[4998]: I1203 17:46:12.487466 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-c4vg8_d93c0246-c19d-45b5-bee1-549fdf19cacd/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 17:46:12 crc kubenswrapper[4998]: I1203 17:46:12.500287 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-77c59f659f-jkw4x_3aa7afc0-1cf1-4e6a-ae4e-98b13411094c/neutron-httpd/0.log" Dec 03 17:46:12 crc kubenswrapper[4998]: I1203 17:46:12.542978 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-77c59f659f-jkw4x_3aa7afc0-1cf1-4e6a-ae4e-98b13411094c/neutron-api/0.log" Dec 03 17:46:13 crc kubenswrapper[4998]: I1203 17:46:13.160966 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_97d1fa21-8374-4329-ac07-27d18803c2ab/nova-cell0-conductor-conductor/0.log" Dec 03 17:46:13 crc kubenswrapper[4998]: I1203 17:46:13.481710 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_2a0de9df-91b8-4cdb-a6a0-46d7f8539b1b/nova-cell1-conductor-conductor/0.log" Dec 03 17:46:13 crc kubenswrapper[4998]: I1203 17:46:13.814945 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_e064512c-fcb1-4c90-8675-bc9f3991c937/nova-cell1-novncproxy-novncproxy/0.log" Dec 03 17:46:14 crc kubenswrapper[4998]: I1203 17:46:14.122637 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-w6l7l_5af60798-c3fe-4b82-9732-4cfd4c442549/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 17:46:14 crc kubenswrapper[4998]: I1203 17:46:14.167815 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_6468949f-85bc-4274-ada0-b2ef4adb0562/nova-api-log/0.log" Dec 03 17:46:14 crc kubenswrapper[4998]: I1203 17:46:14.335193 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_ff8e487a-c00a-47a8-997f-6a0376aa873d/nova-metadata-log/0.log" Dec 03 17:46:14 crc kubenswrapper[4998]: I1203 17:46:14.632319 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_6468949f-85bc-4274-ada0-b2ef4adb0562/nova-api-api/0.log" Dec 03 17:46:14 crc kubenswrapper[4998]: I1203 17:46:14.826997 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_63806609-762c-4623-b4e3-55b90b912b5f/nova-scheduler-scheduler/0.log" Dec 03 17:46:14 crc kubenswrapper[4998]: I1203 17:46:14.887961 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1b2c3df0-703c-438e-9887-cf29844b9eb7/mysql-bootstrap/0.log" Dec 03 17:46:15 crc kubenswrapper[4998]: I1203 17:46:15.020983 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1b2c3df0-703c-438e-9887-cf29844b9eb7/mysql-bootstrap/0.log" Dec 03 17:46:15 crc kubenswrapper[4998]: I1203 17:46:15.081545 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1b2c3df0-703c-438e-9887-cf29844b9eb7/galera/0.log" Dec 03 17:46:15 crc kubenswrapper[4998]: I1203 17:46:15.260648 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8ea3f8b4-86fc-4888-b9c0-36859ab25724/mysql-bootstrap/0.log" Dec 03 17:46:15 crc kubenswrapper[4998]: I1203 17:46:15.452623 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8ea3f8b4-86fc-4888-b9c0-36859ab25724/mysql-bootstrap/0.log" Dec 03 17:46:15 crc kubenswrapper[4998]: I1203 17:46:15.490020 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_8ea3f8b4-86fc-4888-b9c0-36859ab25724/galera/0.log" Dec 03 17:46:15 crc kubenswrapper[4998]: I1203 17:46:15.652201 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_041471bf-1760-4fb8-ac9a-81dc1dd7defa/openstackclient/0.log" Dec 03 17:46:15 crc kubenswrapper[4998]: I1203 17:46:15.829627 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-7sxbw_95a7d4aa-3f41-4d6c-a6ab-3354407e2c2f/ovn-controller/0.log" Dec 03 17:46:15 crc kubenswrapper[4998]: I1203 17:46:15.936382 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-w7sx4_ebaa6f1f-170e-41a4-ad4d-332108a0233b/openstack-network-exporter/0.log" Dec 03 17:46:16 crc kubenswrapper[4998]: I1203 17:46:16.191635 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5hxf2_832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5/ovsdb-server-init/0.log" Dec 03 17:46:16 crc kubenswrapper[4998]: I1203 17:46:16.423300 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5hxf2_832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5/ovsdb-server-init/0.log" Dec 03 17:46:16 crc kubenswrapper[4998]: I1203 17:46:16.430856 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5hxf2_832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5/ovsdb-server/0.log" Dec 03 17:46:16 crc kubenswrapper[4998]: I1203 17:46:16.678828 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-lbwsk_6b9ea60c-d8a5-47a8-91dd-996ee873d6df/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 17:46:16 crc kubenswrapper[4998]: I1203 17:46:16.820040 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-5hxf2_832cc7cb-044b-4bb9-84d9-74e3d5a8f4a5/ovs-vswitchd/0.log" Dec 03 17:46:16 crc kubenswrapper[4998]: I1203 17:46:16.848107 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_ff8e487a-c00a-47a8-997f-6a0376aa873d/nova-metadata-metadata/0.log" Dec 03 17:46:16 crc kubenswrapper[4998]: I1203 17:46:16.880130 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4e93efdf-68ff-4d81-b6b3-447a482e74ff/openstack-network-exporter/0.log" Dec 03 17:46:17 crc kubenswrapper[4998]: I1203 17:46:17.045731 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4e93efdf-68ff-4d81-b6b3-447a482e74ff/ovn-northd/0.log" Dec 03 17:46:17 crc kubenswrapper[4998]: I1203 17:46:17.116419 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_893cf12f-77a9-463f-8dd4-1b3f05b2c9bf/openstack-network-exporter/0.log" Dec 03 17:46:17 crc kubenswrapper[4998]: I1203 17:46:17.260094 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_893cf12f-77a9-463f-8dd4-1b3f05b2c9bf/ovsdbserver-nb/0.log" Dec 03 17:46:17 crc kubenswrapper[4998]: I1203 17:46:17.283981 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_37aa6aee-15f8-4e23-a18c-216b85e605cf/openstack-network-exporter/0.log" Dec 03 17:46:17 crc kubenswrapper[4998]: I1203 17:46:17.359800 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_37aa6aee-15f8-4e23-a18c-216b85e605cf/ovsdbserver-sb/0.log" Dec 03 17:46:17 crc kubenswrapper[4998]: I1203 17:46:17.724499 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-57c877d99d-7qbcv_050d52db-a22f-43de-bf07-ed0b6bed49db/placement-api/0.log" Dec 03 17:46:17 crc kubenswrapper[4998]: I1203 17:46:17.757913 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1/init-config-reloader/0.log" Dec 03 17:46:17 crc kubenswrapper[4998]: I1203 17:46:17.860314 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-57c877d99d-7qbcv_050d52db-a22f-43de-bf07-ed0b6bed49db/placement-log/0.log" Dec 03 17:46:17 crc kubenswrapper[4998]: I1203 17:46:17.903860 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1/init-config-reloader/0.log" Dec 03 17:46:17 crc kubenswrapper[4998]: I1203 17:46:17.925112 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1/config-reloader/0.log" Dec 03 17:46:17 crc kubenswrapper[4998]: I1203 17:46:17.975232 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1/prometheus/0.log" Dec 03 17:46:18 crc kubenswrapper[4998]: I1203 17:46:18.106665 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_b7ad995f-2b7a-4d82-86fa-6ec4528cdeb1/thanos-sidecar/0.log" Dec 03 17:46:18 crc kubenswrapper[4998]: I1203 17:46:18.204777 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_887e67c1-bf49-4540-b448-ce1f3ae5a5f2/setup-container/0.log" Dec 03 17:46:18 crc kubenswrapper[4998]: I1203 17:46:18.449960 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_887e67c1-bf49-4540-b448-ce1f3ae5a5f2/rabbitmq/0.log" Dec 03 17:46:18 crc kubenswrapper[4998]: I1203 17:46:18.472104 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_887e67c1-bf49-4540-b448-ce1f3ae5a5f2/setup-container/0.log" Dec 03 17:46:18 crc kubenswrapper[4998]: I1203 17:46:18.565601 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_dbde2166-b545-4979-845a-8932845b9cd1/setup-container/0.log" Dec 03 17:46:18 crc kubenswrapper[4998]: I1203 17:46:18.797363 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_dbde2166-b545-4979-845a-8932845b9cd1/rabbitmq/0.log" Dec 03 17:46:18 crc kubenswrapper[4998]: I1203 17:46:18.830027 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-notifications-server-0_dbde2166-b545-4979-845a-8932845b9cd1/setup-container/0.log" Dec 03 17:46:18 crc kubenswrapper[4998]: I1203 17:46:18.836927 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_07b66b88-1162-444d-88d7-13dba5be3ce7/setup-container/0.log" Dec 03 17:46:19 crc kubenswrapper[4998]: I1203 17:46:19.041070 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_07b66b88-1162-444d-88d7-13dba5be3ce7/setup-container/0.log" Dec 03 17:46:19 crc kubenswrapper[4998]: I1203 17:46:19.053719 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_07b66b88-1162-444d-88d7-13dba5be3ce7/rabbitmq/0.log" Dec 03 17:46:19 crc kubenswrapper[4998]: I1203 17:46:19.096362 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-8gz2s_d5c35764-7c7b-4346-b54b-6265ec6bca06/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 17:46:19 crc kubenswrapper[4998]: I1203 17:46:19.223393 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-fbwwm_5311a655-a897-486e-80ce-10a0b8082918/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 17:46:19 crc kubenswrapper[4998]: I1203 17:46:19.378910 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-v7f7k_a43a3b83-302e-44ea-9d16-1c0284e8c90f/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 17:46:19 crc kubenswrapper[4998]: I1203 17:46:19.533391 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-29tst_c46460e9-7d52-435d-b5fd-2ebb81570dcf/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 17:46:19 crc kubenswrapper[4998]: I1203 17:46:19.645380 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-97zsr_67c06662-6c0c-484f-b0b9-c61311305323/ssh-known-hosts-edpm-deployment/0.log" Dec 03 17:46:19 crc kubenswrapper[4998]: I1203 17:46:19.870619 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-59d5c55c95-qdh4t_cc160003-2b7c-4772-a0f3-7b431c501fb8/proxy-server/0.log" Dec 03 17:46:19 crc kubenswrapper[4998]: I1203 17:46:19.931479 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-xn4z8_78d5a8fb-51f1-464f-ae8a-f5ecd2be7a0c/swift-ring-rebalance/0.log" Dec 03 17:46:20 crc kubenswrapper[4998]: I1203 17:46:20.077674 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-59d5c55c95-qdh4t_cc160003-2b7c-4772-a0f3-7b431c501fb8/proxy-httpd/0.log" Dec 03 17:46:20 crc kubenswrapper[4998]: I1203 17:46:20.338027 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d01e9a78-9e2c-44c5-9602-4c8860123f88/account-reaper/0.log" Dec 03 17:46:20 crc kubenswrapper[4998]: I1203 17:46:20.362615 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d01e9a78-9e2c-44c5-9602-4c8860123f88/account-auditor/0.log" Dec 03 17:46:20 crc kubenswrapper[4998]: I1203 17:46:20.433833 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d01e9a78-9e2c-44c5-9602-4c8860123f88/account-replicator/0.log" Dec 03 17:46:20 crc kubenswrapper[4998]: I1203 17:46:20.532507 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d01e9a78-9e2c-44c5-9602-4c8860123f88/account-server/0.log" Dec 03 17:46:20 crc kubenswrapper[4998]: I1203 17:46:20.593450 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d01e9a78-9e2c-44c5-9602-4c8860123f88/container-auditor/0.log" Dec 03 17:46:20 crc kubenswrapper[4998]: I1203 17:46:20.654770 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d01e9a78-9e2c-44c5-9602-4c8860123f88/container-replicator/0.log" Dec 03 17:46:20 crc kubenswrapper[4998]: I1203 17:46:20.675499 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d01e9a78-9e2c-44c5-9602-4c8860123f88/container-server/0.log" Dec 03 17:46:20 crc kubenswrapper[4998]: I1203 17:46:20.738533 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d01e9a78-9e2c-44c5-9602-4c8860123f88/container-updater/0.log" Dec 03 17:46:20 crc kubenswrapper[4998]: I1203 17:46:20.746392 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_75024598-618b-49e3-919c-3bf92e4251a2/memcached/0.log" Dec 03 17:46:20 crc kubenswrapper[4998]: I1203 17:46:20.791602 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d01e9a78-9e2c-44c5-9602-4c8860123f88/object-auditor/0.log" Dec 03 17:46:20 crc kubenswrapper[4998]: I1203 17:46:20.859414 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d01e9a78-9e2c-44c5-9602-4c8860123f88/object-expirer/0.log" Dec 03 17:46:20 crc kubenswrapper[4998]: I1203 17:46:20.909800 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d01e9a78-9e2c-44c5-9602-4c8860123f88/object-replicator/0.log" Dec 03 17:46:20 crc kubenswrapper[4998]: I1203 17:46:20.913799 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d01e9a78-9e2c-44c5-9602-4c8860123f88/object-server/0.log" Dec 03 17:46:20 crc kubenswrapper[4998]: I1203 17:46:20.942030 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d01e9a78-9e2c-44c5-9602-4c8860123f88/object-updater/0.log" Dec 03 17:46:20 crc kubenswrapper[4998]: I1203 17:46:20.997494 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d01e9a78-9e2c-44c5-9602-4c8860123f88/rsync/0.log" Dec 03 17:46:21 crc kubenswrapper[4998]: I1203 17:46:21.054810 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d01e9a78-9e2c-44c5-9602-4c8860123f88/swift-recon-cron/0.log" Dec 03 17:46:21 crc kubenswrapper[4998]: I1203 17:46:21.143325 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-t2xc2_10d74042-5cd9-4733-9612-2357a2577f0a/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 17:46:21 crc kubenswrapper[4998]: I1203 17:46:21.387857 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_eb306f4f-0f10-48e1-b20d-78b444da8b8e/test-operator-logs-container/0.log" Dec 03 17:46:21 crc kubenswrapper[4998]: I1203 17:46:21.504638 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_27edc0ce-edf8-4c0e-bd5f-79e58650eed1/tempest-tests-tempest-tests-runner/0.log" Dec 03 17:46:21 crc kubenswrapper[4998]: I1203 17:46:21.550648 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-bw8bd_b407bc5f-1bea-45f6-916a-db593405aab2/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 17:46:22 crc kubenswrapper[4998]: I1203 17:46:22.290122 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-applier-0_880abec3-6fe5-41e0-97c0-43ef4b8aedfe/watcher-applier/0.log" Dec 03 17:46:22 crc kubenswrapper[4998]: I1203 17:46:22.834108 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_baf35f82-8685-4b68-b525-511a2a75aa27/watcher-api-log/0.log" Dec 03 17:46:25 crc kubenswrapper[4998]: I1203 17:46:25.021116 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-decision-engine-0_aaa74904-2894-4cd4-9c3d-6080c69664db/watcher-decision-engine/0.log" Dec 03 17:46:25 crc kubenswrapper[4998]: I1203 17:46:25.972258 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_baf35f82-8685-4b68-b525-511a2a75aa27/watcher-api/0.log" Dec 03 17:46:49 crc kubenswrapper[4998]: I1203 17:46:49.491033 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl_42863f48-3fe1-4510-a729-cd3c0afa39ec/util/0.log" Dec 03 17:46:49 crc kubenswrapper[4998]: I1203 17:46:49.718861 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl_42863f48-3fe1-4510-a729-cd3c0afa39ec/util/0.log" Dec 03 17:46:49 crc kubenswrapper[4998]: I1203 17:46:49.729911 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl_42863f48-3fe1-4510-a729-cd3c0afa39ec/pull/0.log" Dec 03 17:46:49 crc kubenswrapper[4998]: I1203 17:46:49.739927 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl_42863f48-3fe1-4510-a729-cd3c0afa39ec/pull/0.log" Dec 03 17:46:49 crc kubenswrapper[4998]: I1203 17:46:49.920602 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl_42863f48-3fe1-4510-a729-cd3c0afa39ec/pull/0.log" Dec 03 17:46:49 crc kubenswrapper[4998]: I1203 17:46:49.961180 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl_42863f48-3fe1-4510-a729-cd3c0afa39ec/extract/0.log" Dec 03 17:46:49 crc kubenswrapper[4998]: I1203 17:46:49.964807 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_6b154031cfdcc488bebd9cdc3eccafea77107562ca7a14568fb5abe0cc7jdcl_42863f48-3fe1-4510-a729-cd3c0afa39ec/util/0.log" Dec 03 17:46:50 crc kubenswrapper[4998]: I1203 17:46:50.167486 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-5h65w_da7a11af-1959-4c47-96d3-4c5f4c469493/kube-rbac-proxy/0.log" Dec 03 17:46:50 crc kubenswrapper[4998]: I1203 17:46:50.218125 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-tv6bd_a67cb778-507e-4dba-9ae4-60ae60315a0a/kube-rbac-proxy/0.log" Dec 03 17:46:50 crc kubenswrapper[4998]: I1203 17:46:50.253588 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-5h65w_da7a11af-1959-4c47-96d3-4c5f4c469493/manager/0.log" Dec 03 17:46:50 crc kubenswrapper[4998]: I1203 17:46:50.403411 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-m4gc7_67e57367-12f9-487d-b691-ca243a28b056/kube-rbac-proxy/0.log" Dec 03 17:46:50 crc kubenswrapper[4998]: I1203 17:46:50.419932 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-tv6bd_a67cb778-507e-4dba-9ae4-60ae60315a0a/manager/0.log" Dec 03 17:46:50 crc kubenswrapper[4998]: I1203 17:46:50.469667 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-m4gc7_67e57367-12f9-487d-b691-ca243a28b056/manager/0.log" Dec 03 17:46:50 crc kubenswrapper[4998]: I1203 17:46:50.600742 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-tkz2j_624bac01-c09f-424f-9000-feb138765cc4/kube-rbac-proxy/0.log" Dec 03 17:46:50 crc kubenswrapper[4998]: I1203 17:46:50.651921 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-tkz2j_624bac01-c09f-424f-9000-feb138765cc4/manager/0.log" Dec 03 17:46:50 crc kubenswrapper[4998]: I1203 17:46:50.759587 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-d2gcr_e52d7801-a5c9-47a3-bedc-27174cfbd993/manager/0.log" Dec 03 17:46:50 crc kubenswrapper[4998]: I1203 17:46:50.781045 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-d2gcr_e52d7801-a5c9-47a3-bedc-27174cfbd993/kube-rbac-proxy/0.log" Dec 03 17:46:50 crc kubenswrapper[4998]: I1203 17:46:50.865036 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-gzg25_7c372d0c-4cac-4e4c-bc8f-409b5c151d53/kube-rbac-proxy/0.log" Dec 03 17:46:50 crc kubenswrapper[4998]: I1203 17:46:50.982868 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-gzg25_7c372d0c-4cac-4e4c-bc8f-409b5c151d53/manager/0.log" Dec 03 17:46:51 crc kubenswrapper[4998]: I1203 17:46:51.042744 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-hr74g_ade59116-1aaf-44e0-b47b-15bdc7b402f0/kube-rbac-proxy/0.log" Dec 03 17:46:51 crc kubenswrapper[4998]: I1203 17:46:51.248489 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-hr74g_ade59116-1aaf-44e0-b47b-15bdc7b402f0/manager/0.log" Dec 03 17:46:51 crc kubenswrapper[4998]: I1203 17:46:51.248743 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-tcltm_17cdd58c-7a0c-42e0-bcbe-4545050637e3/manager/0.log" Dec 03 17:46:51 crc kubenswrapper[4998]: I1203 17:46:51.284844 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-tcltm_17cdd58c-7a0c-42e0-bcbe-4545050637e3/kube-rbac-proxy/0.log" Dec 03 17:46:51 crc kubenswrapper[4998]: I1203 17:46:51.523992 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-nfm6p_dcb7317b-497a-4135-be25-3462a91406e9/kube-rbac-proxy/0.log" Dec 03 17:46:51 crc kubenswrapper[4998]: I1203 17:46:51.543859 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-nfm6p_dcb7317b-497a-4135-be25-3462a91406e9/manager/0.log" Dec 03 17:46:51 crc kubenswrapper[4998]: I1203 17:46:51.646810 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-mn69l_c92de5b9-8ef5-4f4c-b881-82c5f05cd853/kube-rbac-proxy/0.log" Dec 03 17:46:51 crc kubenswrapper[4998]: I1203 17:46:51.686112 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-mn69l_c92de5b9-8ef5-4f4c-b881-82c5f05cd853/manager/0.log" Dec 03 17:46:51 crc kubenswrapper[4998]: I1203 17:46:51.744324 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-f4jqc_0c9761a8-29fb-4d99-9bc0-914a041b36ac/kube-rbac-proxy/0.log" Dec 03 17:46:51 crc kubenswrapper[4998]: I1203 17:46:51.884179 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-f4jqc_0c9761a8-29fb-4d99-9bc0-914a041b36ac/manager/0.log" Dec 03 17:46:51 crc kubenswrapper[4998]: I1203 17:46:51.943818 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-hzl7f_0e92e441-8150-4817-8039-9cdc8424e562/kube-rbac-proxy/0.log" Dec 03 17:46:52 crc kubenswrapper[4998]: I1203 17:46:52.007452 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-hzl7f_0e92e441-8150-4817-8039-9cdc8424e562/manager/0.log" Dec 03 17:46:52 crc kubenswrapper[4998]: I1203 17:46:52.762686 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-2h2m5_f23eb2b1-107d-4167-a695-4479df0ae92a/kube-rbac-proxy/0.log" Dec 03 17:46:52 crc kubenswrapper[4998]: I1203 17:46:52.854274 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-2h2m5_f23eb2b1-107d-4167-a695-4479df0ae92a/manager/0.log" Dec 03 17:46:52 crc kubenswrapper[4998]: I1203 17:46:52.876714 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-ph8r8_d926336a-658f-49bd-81b9-8086c8bb97a5/kube-rbac-proxy/0.log" Dec 03 17:46:52 crc kubenswrapper[4998]: I1203 17:46:52.993208 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-ph8r8_d926336a-658f-49bd-81b9-8086c8bb97a5/manager/0.log" Dec 03 17:46:53 crc kubenswrapper[4998]: I1203 17:46:53.066804 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz_42a2439a-98ee-4172-813c-9ce7109fb1c0/kube-rbac-proxy/0.log" Dec 03 17:46:53 crc kubenswrapper[4998]: I1203 17:46:53.100462 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd44fnhz_42a2439a-98ee-4172-813c-9ce7109fb1c0/manager/0.log" Dec 03 17:46:53 crc kubenswrapper[4998]: I1203 17:46:53.530976 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-59449b5fc9-lss2t_f730843a-f6a7-4e56-bd88-333d2531d2f0/operator/0.log" Dec 03 17:46:53 crc kubenswrapper[4998]: I1203 17:46:53.794490 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-8rmlq_3e42935d-e3d2-48a6-a787-caa1ed1f7ef1/registry-server/0.log" Dec 03 17:46:53 crc kubenswrapper[4998]: I1203 17:46:53.911518 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-28xct_79a1881c-77c7-4c5c-b2fc-b2d7d9c31496/kube-rbac-proxy/0.log" Dec 03 17:46:54 crc kubenswrapper[4998]: I1203 17:46:54.051361 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-28xct_79a1881c-77c7-4c5c-b2fc-b2d7d9c31496/manager/0.log" Dec 03 17:46:54 crc kubenswrapper[4998]: I1203 17:46:54.114871 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-h72w5_77fb606c-dea3-4e34-9f4c-5bb305e3b7f4/kube-rbac-proxy/0.log" Dec 03 17:46:54 crc kubenswrapper[4998]: I1203 17:46:54.231128 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-h72w5_77fb606c-dea3-4e34-9f4c-5bb305e3b7f4/manager/0.log" Dec 03 17:46:54 crc kubenswrapper[4998]: I1203 17:46:54.428631 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-fqdqg_9f587468-9ae1-43af-8652-f4394865ff37/operator/0.log" Dec 03 17:46:54 crc kubenswrapper[4998]: I1203 17:46:54.434778 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-2wrqc_d7adeced-3036-49a7-8806-fdbefce6406f/kube-rbac-proxy/0.log" Dec 03 17:46:54 crc kubenswrapper[4998]: I1203 17:46:54.547145 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-67847ddbd8-t8wbd_07d50ab9-95b8-4c58-9192-445ec4e32b59/manager/0.log" Dec 03 17:46:54 crc kubenswrapper[4998]: I1203 17:46:54.595529 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-2wrqc_d7adeced-3036-49a7-8806-fdbefce6406f/manager/0.log" Dec 03 17:46:54 crc kubenswrapper[4998]: I1203 17:46:54.680141 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-f6vdv_aa3ac0c7-ea42-4728-8406-0ce3f9a96c21/kube-rbac-proxy/0.log" Dec 03 17:46:54 crc kubenswrapper[4998]: I1203 17:46:54.807418 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-f6vdv_aa3ac0c7-ea42-4728-8406-0ce3f9a96c21/manager/0.log" Dec 03 17:46:54 crc kubenswrapper[4998]: I1203 17:46:54.839699 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-z8g4g_6b89567b-4963-4113-9900-a132d873d260/kube-rbac-proxy/0.log" Dec 03 17:46:54 crc kubenswrapper[4998]: I1203 17:46:54.906144 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-z8g4g_6b89567b-4963-4113-9900-a132d873d260/manager/0.log" Dec 03 17:46:55 crc kubenswrapper[4998]: I1203 17:46:55.055111 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-f77f967dc-ndj9c_ac213ee6-009b-4098-9d18-6dfc8cb97a0e/kube-rbac-proxy/0.log" Dec 03 17:46:55 crc kubenswrapper[4998]: I1203 17:46:55.070789 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-f77f967dc-ndj9c_ac213ee6-009b-4098-9d18-6dfc8cb97a0e/manager/0.log" Dec 03 17:47:03 crc kubenswrapper[4998]: I1203 17:47:03.333637 4998 scope.go:117] "RemoveContainer" containerID="860280a2225d4ab31ba7d1d05349206c0f6479a01e7da2c97da56330502c002b" Dec 03 17:47:03 crc kubenswrapper[4998]: I1203 17:47:03.366397 4998 scope.go:117] "RemoveContainer" containerID="713870b960eed30b68ad1f1b26b8be535d97a2bf63192376a07cd4560b924b30" Dec 03 17:47:03 crc kubenswrapper[4998]: I1203 17:47:03.434051 4998 scope.go:117] "RemoveContainer" containerID="7c7d717b1e29083a433e71613e0683a5404c9b376452940760f7722e834726ff" Dec 03 17:47:15 crc kubenswrapper[4998]: I1203 17:47:15.133382 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-sn2r4_12d866d2-1002-4db4-9832-59b5fea5062e/control-plane-machine-set-operator/0.log" Dec 03 17:47:15 crc kubenswrapper[4998]: I1203 17:47:15.349884 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-ggw6k_892e64fc-db6b-4410-9b4c-3ff78cc82074/kube-rbac-proxy/0.log" Dec 03 17:47:15 crc kubenswrapper[4998]: I1203 17:47:15.355965 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-ggw6k_892e64fc-db6b-4410-9b4c-3ff78cc82074/machine-api-operator/0.log" Dec 03 17:47:27 crc kubenswrapper[4998]: I1203 17:47:27.110782 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:47:27 crc kubenswrapper[4998]: I1203 17:47:27.111343 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:47:29 crc kubenswrapper[4998]: I1203 17:47:29.029107 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-7sddv_ab951659-919a-4a60-b494-56278d2cd184/cert-manager-controller/0.log" Dec 03 17:47:29 crc kubenswrapper[4998]: I1203 17:47:29.114115 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-p7n29_b12a5054-c73c-4d99-a5a8-174decd7d642/cert-manager-cainjector/0.log" Dec 03 17:47:29 crc kubenswrapper[4998]: I1203 17:47:29.265076 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-xcmb6_f4bb9920-07e2-4f7d-a81b-e91c374e8f28/cert-manager-webhook/0.log" Dec 03 17:47:42 crc kubenswrapper[4998]: I1203 17:47:42.054153 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-5td8z_9d6e2533-1816-4ea6-a1cc-dc7f834e1c33/nmstate-console-plugin/0.log" Dec 03 17:47:42 crc kubenswrapper[4998]: I1203 17:47:42.229356 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-plkwj_4a388be2-7ba3-4fdd-896f-28577a929261/nmstate-handler/0.log" Dec 03 17:47:42 crc kubenswrapper[4998]: I1203 17:47:42.254719 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-n9fkd_c74a341f-037c-472d-af3c-d76406b0f422/kube-rbac-proxy/0.log" Dec 03 17:47:42 crc kubenswrapper[4998]: I1203 17:47:42.292368 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-n9fkd_c74a341f-037c-472d-af3c-d76406b0f422/nmstate-metrics/0.log" Dec 03 17:47:42 crc kubenswrapper[4998]: I1203 17:47:42.452315 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-7rcw5_bf4a690b-a2ab-41b9-a6d6-f053d2c4d735/nmstate-webhook/0.log" Dec 03 17:47:42 crc kubenswrapper[4998]: I1203 17:47:42.484812 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-cdghz_2f9b5504-1b4e-4ca7-8fc7-1319edc59ce9/nmstate-operator/0.log" Dec 03 17:47:57 crc kubenswrapper[4998]: I1203 17:47:57.111084 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:47:57 crc kubenswrapper[4998]: I1203 17:47:57.113017 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:47:57 crc kubenswrapper[4998]: I1203 17:47:57.189362 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-d9v5x_9ce7261b-c84b-473e-a501-b570a4ba5b02/kube-rbac-proxy/0.log" Dec 03 17:47:57 crc kubenswrapper[4998]: I1203 17:47:57.340724 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-d9v5x_9ce7261b-c84b-473e-a501-b570a4ba5b02/controller/0.log" Dec 03 17:47:57 crc kubenswrapper[4998]: I1203 17:47:57.395502 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-m449d_45a612f3-9622-4ee1-be7e-04e7c5035f7a/cp-frr-files/0.log" Dec 03 17:47:57 crc kubenswrapper[4998]: I1203 17:47:57.602847 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-m449d_45a612f3-9622-4ee1-be7e-04e7c5035f7a/cp-frr-files/0.log" Dec 03 17:47:57 crc kubenswrapper[4998]: I1203 17:47:57.670850 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-m449d_45a612f3-9622-4ee1-be7e-04e7c5035f7a/cp-metrics/0.log" Dec 03 17:47:57 crc kubenswrapper[4998]: I1203 17:47:57.670932 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-m449d_45a612f3-9622-4ee1-be7e-04e7c5035f7a/cp-reloader/0.log" Dec 03 17:47:57 crc kubenswrapper[4998]: I1203 17:47:57.691742 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-m449d_45a612f3-9622-4ee1-be7e-04e7c5035f7a/cp-reloader/0.log" Dec 03 17:47:57 crc kubenswrapper[4998]: I1203 17:47:57.881608 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-m449d_45a612f3-9622-4ee1-be7e-04e7c5035f7a/cp-frr-files/0.log" Dec 03 17:47:57 crc kubenswrapper[4998]: I1203 17:47:57.908166 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-m449d_45a612f3-9622-4ee1-be7e-04e7c5035f7a/cp-reloader/0.log" Dec 03 17:47:57 crc kubenswrapper[4998]: I1203 17:47:57.912083 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-m449d_45a612f3-9622-4ee1-be7e-04e7c5035f7a/cp-metrics/0.log" Dec 03 17:47:57 crc kubenswrapper[4998]: I1203 17:47:57.954554 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-m449d_45a612f3-9622-4ee1-be7e-04e7c5035f7a/cp-metrics/0.log" Dec 03 17:47:58 crc kubenswrapper[4998]: I1203 17:47:58.162828 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-m449d_45a612f3-9622-4ee1-be7e-04e7c5035f7a/cp-metrics/0.log" Dec 03 17:47:58 crc kubenswrapper[4998]: I1203 17:47:58.172094 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-m449d_45a612f3-9622-4ee1-be7e-04e7c5035f7a/cp-frr-files/0.log" Dec 03 17:47:58 crc kubenswrapper[4998]: I1203 17:47:58.191269 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-m449d_45a612f3-9622-4ee1-be7e-04e7c5035f7a/cp-reloader/0.log" Dec 03 17:47:58 crc kubenswrapper[4998]: I1203 17:47:58.216238 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-m449d_45a612f3-9622-4ee1-be7e-04e7c5035f7a/controller/0.log" Dec 03 17:47:58 crc kubenswrapper[4998]: I1203 17:47:58.373872 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-m449d_45a612f3-9622-4ee1-be7e-04e7c5035f7a/kube-rbac-proxy/0.log" Dec 03 17:47:58 crc kubenswrapper[4998]: I1203 17:47:58.382678 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-m449d_45a612f3-9622-4ee1-be7e-04e7c5035f7a/frr-metrics/0.log" Dec 03 17:47:58 crc kubenswrapper[4998]: I1203 17:47:58.438521 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-m449d_45a612f3-9622-4ee1-be7e-04e7c5035f7a/kube-rbac-proxy-frr/0.log" Dec 03 17:47:58 crc kubenswrapper[4998]: I1203 17:47:58.558228 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-m449d_45a612f3-9622-4ee1-be7e-04e7c5035f7a/reloader/0.log" Dec 03 17:47:58 crc kubenswrapper[4998]: I1203 17:47:58.696836 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-j8sl8_cdd57e8f-0d60-4634-8460-2bee971c0bc7/frr-k8s-webhook-server/0.log" Dec 03 17:47:58 crc kubenswrapper[4998]: I1203 17:47:58.894834 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5d9448b85b-kszzj_a4f19703-b458-4f44-b824-6c249ccf6194/manager/0.log" Dec 03 17:47:59 crc kubenswrapper[4998]: I1203 17:47:59.052667 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-f5b585465-glgcb_40a7ef48-e7b8-4668-8216-bb6392862991/webhook-server/0.log" Dec 03 17:47:59 crc kubenswrapper[4998]: I1203 17:47:59.106789 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qd4fq_3830057a-341e-4586-8f1a-f80d34f8c997/kube-rbac-proxy/0.log" Dec 03 17:47:59 crc kubenswrapper[4998]: I1203 17:47:59.811262 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qd4fq_3830057a-341e-4586-8f1a-f80d34f8c997/speaker/0.log" Dec 03 17:48:00 crc kubenswrapper[4998]: I1203 17:48:00.078661 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-m449d_45a612f3-9622-4ee1-be7e-04e7c5035f7a/frr/0.log" Dec 03 17:48:13 crc kubenswrapper[4998]: I1203 17:48:13.063183 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l_b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99/util/0.log" Dec 03 17:48:13 crc kubenswrapper[4998]: I1203 17:48:13.314099 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l_b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99/util/0.log" Dec 03 17:48:13 crc kubenswrapper[4998]: I1203 17:48:13.325300 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l_b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99/pull/0.log" Dec 03 17:48:13 crc kubenswrapper[4998]: I1203 17:48:13.368223 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l_b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99/pull/0.log" Dec 03 17:48:13 crc kubenswrapper[4998]: I1203 17:48:13.508115 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l_b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99/util/0.log" Dec 03 17:48:13 crc kubenswrapper[4998]: I1203 17:48:13.527956 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l_b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99/pull/0.log" Dec 03 17:48:13 crc kubenswrapper[4998]: I1203 17:48:13.553976 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqpm6l_b7bba5b4-55d7-465f-89a3-3ed3b1cd0d99/extract/0.log" Dec 03 17:48:13 crc kubenswrapper[4998]: I1203 17:48:13.671458 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj_9038bb91-6c7d-42c6-b864-9d522c73c488/util/0.log" Dec 03 17:48:13 crc kubenswrapper[4998]: I1203 17:48:13.865838 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj_9038bb91-6c7d-42c6-b864-9d522c73c488/pull/0.log" Dec 03 17:48:13 crc kubenswrapper[4998]: I1203 17:48:13.907212 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj_9038bb91-6c7d-42c6-b864-9d522c73c488/pull/0.log" Dec 03 17:48:13 crc kubenswrapper[4998]: I1203 17:48:13.912529 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj_9038bb91-6c7d-42c6-b864-9d522c73c488/util/0.log" Dec 03 17:48:14 crc kubenswrapper[4998]: I1203 17:48:14.096472 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj_9038bb91-6c7d-42c6-b864-9d522c73c488/util/0.log" Dec 03 17:48:14 crc kubenswrapper[4998]: I1203 17:48:14.099531 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj_9038bb91-6c7d-42c6-b864-9d522c73c488/pull/0.log" Dec 03 17:48:14 crc kubenswrapper[4998]: I1203 17:48:14.108103 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92109r7bj_9038bb91-6c7d-42c6-b864-9d522c73c488/extract/0.log" Dec 03 17:48:14 crc kubenswrapper[4998]: I1203 17:48:14.278568 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl_f399c383-c804-4c6e-a247-0c09ca313644/util/0.log" Dec 03 17:48:14 crc kubenswrapper[4998]: I1203 17:48:14.424089 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl_f399c383-c804-4c6e-a247-0c09ca313644/util/0.log" Dec 03 17:48:14 crc kubenswrapper[4998]: I1203 17:48:14.433134 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl_f399c383-c804-4c6e-a247-0c09ca313644/pull/0.log" Dec 03 17:48:14 crc kubenswrapper[4998]: I1203 17:48:14.471647 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl_f399c383-c804-4c6e-a247-0c09ca313644/pull/0.log" Dec 03 17:48:14 crc kubenswrapper[4998]: I1203 17:48:14.664713 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl_f399c383-c804-4c6e-a247-0c09ca313644/util/0.log" Dec 03 17:48:14 crc kubenswrapper[4998]: I1203 17:48:14.666042 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl_f399c383-c804-4c6e-a247-0c09ca313644/pull/0.log" Dec 03 17:48:14 crc kubenswrapper[4998]: I1203 17:48:14.694109 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f8376nhl_f399c383-c804-4c6e-a247-0c09ca313644/extract/0.log" Dec 03 17:48:14 crc kubenswrapper[4998]: I1203 17:48:14.836078 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tdx5r_f0e26a80-b16f-4dd8-a8b1-c567913e3922/extract-utilities/0.log" Dec 03 17:48:15 crc kubenswrapper[4998]: I1203 17:48:15.037943 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tdx5r_f0e26a80-b16f-4dd8-a8b1-c567913e3922/extract-utilities/0.log" Dec 03 17:48:15 crc kubenswrapper[4998]: I1203 17:48:15.048470 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tdx5r_f0e26a80-b16f-4dd8-a8b1-c567913e3922/extract-content/0.log" Dec 03 17:48:15 crc kubenswrapper[4998]: I1203 17:48:15.048617 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tdx5r_f0e26a80-b16f-4dd8-a8b1-c567913e3922/extract-content/0.log" Dec 03 17:48:15 crc kubenswrapper[4998]: I1203 17:48:15.224688 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tdx5r_f0e26a80-b16f-4dd8-a8b1-c567913e3922/extract-utilities/0.log" Dec 03 17:48:15 crc kubenswrapper[4998]: I1203 17:48:15.261313 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tdx5r_f0e26a80-b16f-4dd8-a8b1-c567913e3922/extract-content/0.log" Dec 03 17:48:15 crc kubenswrapper[4998]: I1203 17:48:15.435543 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t6bk2_999ac090-324b-4348-8feb-215ecb6e509b/extract-utilities/0.log" Dec 03 17:48:15 crc kubenswrapper[4998]: I1203 17:48:15.712131 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t6bk2_999ac090-324b-4348-8feb-215ecb6e509b/extract-content/0.log" Dec 03 17:48:15 crc kubenswrapper[4998]: I1203 17:48:15.726311 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t6bk2_999ac090-324b-4348-8feb-215ecb6e509b/extract-utilities/0.log" Dec 03 17:48:15 crc kubenswrapper[4998]: I1203 17:48:15.745646 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t6bk2_999ac090-324b-4348-8feb-215ecb6e509b/extract-content/0.log" Dec 03 17:48:15 crc kubenswrapper[4998]: I1203 17:48:15.965560 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t6bk2_999ac090-324b-4348-8feb-215ecb6e509b/extract-utilities/0.log" Dec 03 17:48:16 crc kubenswrapper[4998]: I1203 17:48:16.129732 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t6bk2_999ac090-324b-4348-8feb-215ecb6e509b/extract-content/0.log" Dec 03 17:48:16 crc kubenswrapper[4998]: I1203 17:48:16.444114 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tdx5r_f0e26a80-b16f-4dd8-a8b1-c567913e3922/registry-server/0.log" Dec 03 17:48:16 crc kubenswrapper[4998]: I1203 17:48:16.530325 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t6bk2_999ac090-324b-4348-8feb-215ecb6e509b/registry-server/0.log" Dec 03 17:48:16 crc kubenswrapper[4998]: I1203 17:48:16.534275 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-ts48f_4d7e0e87-75f9-437a-916a-f702c53d370b/marketplace-operator/0.log" Dec 03 17:48:16 crc kubenswrapper[4998]: I1203 17:48:16.662314 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rnsnx_20cc0757-ee9b-431b-85ea-189a0b33827b/extract-utilities/0.log" Dec 03 17:48:16 crc kubenswrapper[4998]: I1203 17:48:16.818868 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rnsnx_20cc0757-ee9b-431b-85ea-189a0b33827b/extract-content/0.log" Dec 03 17:48:16 crc kubenswrapper[4998]: I1203 17:48:16.823190 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rnsnx_20cc0757-ee9b-431b-85ea-189a0b33827b/extract-utilities/0.log" Dec 03 17:48:16 crc kubenswrapper[4998]: I1203 17:48:16.827814 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rnsnx_20cc0757-ee9b-431b-85ea-189a0b33827b/extract-content/0.log" Dec 03 17:48:16 crc kubenswrapper[4998]: I1203 17:48:16.992401 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rnsnx_20cc0757-ee9b-431b-85ea-189a0b33827b/extract-utilities/0.log" Dec 03 17:48:17 crc kubenswrapper[4998]: I1203 17:48:17.002163 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rnsnx_20cc0757-ee9b-431b-85ea-189a0b33827b/extract-content/0.log" Dec 03 17:48:17 crc kubenswrapper[4998]: I1203 17:48:17.096745 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jn68h_5d1c51e3-2a5d-45dc-917f-0101d813f6d6/extract-utilities/0.log" Dec 03 17:48:17 crc kubenswrapper[4998]: I1203 17:48:17.250890 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rnsnx_20cc0757-ee9b-431b-85ea-189a0b33827b/registry-server/0.log" Dec 03 17:48:17 crc kubenswrapper[4998]: I1203 17:48:17.282988 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jn68h_5d1c51e3-2a5d-45dc-917f-0101d813f6d6/extract-content/0.log" Dec 03 17:48:17 crc kubenswrapper[4998]: I1203 17:48:17.287430 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jn68h_5d1c51e3-2a5d-45dc-917f-0101d813f6d6/extract-utilities/0.log" Dec 03 17:48:17 crc kubenswrapper[4998]: I1203 17:48:17.319673 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jn68h_5d1c51e3-2a5d-45dc-917f-0101d813f6d6/extract-content/0.log" Dec 03 17:48:17 crc kubenswrapper[4998]: I1203 17:48:17.460170 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jn68h_5d1c51e3-2a5d-45dc-917f-0101d813f6d6/extract-utilities/0.log" Dec 03 17:48:17 crc kubenswrapper[4998]: I1203 17:48:17.475346 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jn68h_5d1c51e3-2a5d-45dc-917f-0101d813f6d6/extract-content/0.log" Dec 03 17:48:18 crc kubenswrapper[4998]: I1203 17:48:18.254563 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jn68h_5d1c51e3-2a5d-45dc-917f-0101d813f6d6/registry-server/0.log" Dec 03 17:48:27 crc kubenswrapper[4998]: I1203 17:48:27.111361 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:48:27 crc kubenswrapper[4998]: I1203 17:48:27.111890 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:48:27 crc kubenswrapper[4998]: I1203 17:48:27.111938 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 17:48:27 crc kubenswrapper[4998]: I1203 17:48:27.112803 4998 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315"} pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:48:27 crc kubenswrapper[4998]: I1203 17:48:27.112870 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" containerID="cri-o://09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" gracePeriod=600 Dec 03 17:48:27 crc kubenswrapper[4998]: E1203 17:48:27.253451 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:48:27 crc kubenswrapper[4998]: I1203 17:48:27.724319 4998 generic.go:334] "Generic (PLEG): container finished" podID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" exitCode=0 Dec 03 17:48:27 crc kubenswrapper[4998]: I1203 17:48:27.724387 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerDied","Data":"09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315"} Dec 03 17:48:27 crc kubenswrapper[4998]: I1203 17:48:27.724703 4998 scope.go:117] "RemoveContainer" containerID="c32ad81ca8fbe8808005dc071a77b0419d9a93bdb054bce341dafb8dbf1dd822" Dec 03 17:48:27 crc kubenswrapper[4998]: I1203 17:48:27.725447 4998 scope.go:117] "RemoveContainer" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" Dec 03 17:48:27 crc kubenswrapper[4998]: E1203 17:48:27.725777 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:48:30 crc kubenswrapper[4998]: I1203 17:48:30.709419 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-f8jt6_50b3e173-2cde-4732-9fa4-0d5cc7936dcd/prometheus-operator/0.log" Dec 03 17:48:30 crc kubenswrapper[4998]: I1203 17:48:30.887884 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5c4dbf9785-4ffgs_6a29f86a-9f1e-44b6-a1e2-1cdc7dd3e7ec/prometheus-operator-admission-webhook/0.log" Dec 03 17:48:30 crc kubenswrapper[4998]: I1203 17:48:30.940089 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5c4dbf9785-d46s4_0dedded0-9a7b-4697-9a69-6ca56016f0f1/prometheus-operator-admission-webhook/0.log" Dec 03 17:48:31 crc kubenswrapper[4998]: I1203 17:48:31.091464 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-zzvhx_60293ca8-98bb-49a8-b54f-0c3aa3f84cc3/operator/0.log" Dec 03 17:48:31 crc kubenswrapper[4998]: I1203 17:48:31.158498 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-2wk7t_48acd5a8-2b63-4683-9f39-43f3e3193a85/perses-operator/0.log" Dec 03 17:48:40 crc kubenswrapper[4998]: I1203 17:48:40.678019 4998 scope.go:117] "RemoveContainer" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" Dec 03 17:48:40 crc kubenswrapper[4998]: E1203 17:48:40.678677 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:48:45 crc kubenswrapper[4998]: E1203 17:48:45.880430 4998 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.64:57544->38.102.83.64:36649: write tcp 38.102.83.64:57544->38.102.83.64:36649: write: broken pipe Dec 03 17:48:51 crc kubenswrapper[4998]: I1203 17:48:51.686344 4998 scope.go:117] "RemoveContainer" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" Dec 03 17:48:51 crc kubenswrapper[4998]: E1203 17:48:51.687127 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:49:03 crc kubenswrapper[4998]: I1203 17:49:03.678938 4998 scope.go:117] "RemoveContainer" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" Dec 03 17:49:03 crc kubenswrapper[4998]: E1203 17:49:03.679975 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:49:15 crc kubenswrapper[4998]: I1203 17:49:15.679285 4998 scope.go:117] "RemoveContainer" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" Dec 03 17:49:15 crc kubenswrapper[4998]: E1203 17:49:15.680162 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:49:30 crc kubenswrapper[4998]: I1203 17:49:30.678973 4998 scope.go:117] "RemoveContainer" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" Dec 03 17:49:30 crc kubenswrapper[4998]: E1203 17:49:30.679674 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:49:44 crc kubenswrapper[4998]: I1203 17:49:44.679235 4998 scope.go:117] "RemoveContainer" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" Dec 03 17:49:44 crc kubenswrapper[4998]: E1203 17:49:44.683173 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:49:56 crc kubenswrapper[4998]: I1203 17:49:56.678742 4998 scope.go:117] "RemoveContainer" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" Dec 03 17:49:56 crc kubenswrapper[4998]: E1203 17:49:56.681284 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:50:10 crc kubenswrapper[4998]: I1203 17:50:10.678537 4998 scope.go:117] "RemoveContainer" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" Dec 03 17:50:10 crc kubenswrapper[4998]: E1203 17:50:10.679304 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:50:21 crc kubenswrapper[4998]: I1203 17:50:21.583182 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7bmzz"] Dec 03 17:50:21 crc kubenswrapper[4998]: E1203 17:50:21.586150 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad81770d-9852-4abb-97b6-c803bce55972" containerName="extract-utilities" Dec 03 17:50:21 crc kubenswrapper[4998]: I1203 17:50:21.586336 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad81770d-9852-4abb-97b6-c803bce55972" containerName="extract-utilities" Dec 03 17:50:21 crc kubenswrapper[4998]: E1203 17:50:21.586484 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad81770d-9852-4abb-97b6-c803bce55972" containerName="registry-server" Dec 03 17:50:21 crc kubenswrapper[4998]: I1203 17:50:21.586596 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad81770d-9852-4abb-97b6-c803bce55972" containerName="registry-server" Dec 03 17:50:21 crc kubenswrapper[4998]: E1203 17:50:21.586726 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad81770d-9852-4abb-97b6-c803bce55972" containerName="extract-content" Dec 03 17:50:21 crc kubenswrapper[4998]: I1203 17:50:21.586883 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad81770d-9852-4abb-97b6-c803bce55972" containerName="extract-content" Dec 03 17:50:21 crc kubenswrapper[4998]: I1203 17:50:21.587370 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad81770d-9852-4abb-97b6-c803bce55972" containerName="registry-server" Dec 03 17:50:21 crc kubenswrapper[4998]: I1203 17:50:21.590710 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7bmzz" Dec 03 17:50:21 crc kubenswrapper[4998]: I1203 17:50:21.598517 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7bmzz"] Dec 03 17:50:21 crc kubenswrapper[4998]: I1203 17:50:21.600645 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87kxv\" (UniqueName: \"kubernetes.io/projected/9fe16aa3-7744-4763-869b-a08d233cca19-kube-api-access-87kxv\") pod \"redhat-marketplace-7bmzz\" (UID: \"9fe16aa3-7744-4763-869b-a08d233cca19\") " pod="openshift-marketplace/redhat-marketplace-7bmzz" Dec 03 17:50:21 crc kubenswrapper[4998]: I1203 17:50:21.600802 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fe16aa3-7744-4763-869b-a08d233cca19-catalog-content\") pod \"redhat-marketplace-7bmzz\" (UID: \"9fe16aa3-7744-4763-869b-a08d233cca19\") " pod="openshift-marketplace/redhat-marketplace-7bmzz" Dec 03 17:50:21 crc kubenswrapper[4998]: I1203 17:50:21.600838 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fe16aa3-7744-4763-869b-a08d233cca19-utilities\") pod \"redhat-marketplace-7bmzz\" (UID: \"9fe16aa3-7744-4763-869b-a08d233cca19\") " pod="openshift-marketplace/redhat-marketplace-7bmzz" Dec 03 17:50:21 crc kubenswrapper[4998]: I1203 17:50:21.703546 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fe16aa3-7744-4763-869b-a08d233cca19-catalog-content\") pod \"redhat-marketplace-7bmzz\" (UID: \"9fe16aa3-7744-4763-869b-a08d233cca19\") " pod="openshift-marketplace/redhat-marketplace-7bmzz" Dec 03 17:50:21 crc kubenswrapper[4998]: I1203 17:50:21.703606 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fe16aa3-7744-4763-869b-a08d233cca19-utilities\") pod \"redhat-marketplace-7bmzz\" (UID: \"9fe16aa3-7744-4763-869b-a08d233cca19\") " pod="openshift-marketplace/redhat-marketplace-7bmzz" Dec 03 17:50:21 crc kubenswrapper[4998]: I1203 17:50:21.703726 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87kxv\" (UniqueName: \"kubernetes.io/projected/9fe16aa3-7744-4763-869b-a08d233cca19-kube-api-access-87kxv\") pod \"redhat-marketplace-7bmzz\" (UID: \"9fe16aa3-7744-4763-869b-a08d233cca19\") " pod="openshift-marketplace/redhat-marketplace-7bmzz" Dec 03 17:50:21 crc kubenswrapper[4998]: I1203 17:50:21.704505 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fe16aa3-7744-4763-869b-a08d233cca19-catalog-content\") pod \"redhat-marketplace-7bmzz\" (UID: \"9fe16aa3-7744-4763-869b-a08d233cca19\") " pod="openshift-marketplace/redhat-marketplace-7bmzz" Dec 03 17:50:21 crc kubenswrapper[4998]: I1203 17:50:21.704505 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fe16aa3-7744-4763-869b-a08d233cca19-utilities\") pod \"redhat-marketplace-7bmzz\" (UID: \"9fe16aa3-7744-4763-869b-a08d233cca19\") " pod="openshift-marketplace/redhat-marketplace-7bmzz" Dec 03 17:50:21 crc kubenswrapper[4998]: I1203 17:50:21.727152 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87kxv\" (UniqueName: \"kubernetes.io/projected/9fe16aa3-7744-4763-869b-a08d233cca19-kube-api-access-87kxv\") pod \"redhat-marketplace-7bmzz\" (UID: \"9fe16aa3-7744-4763-869b-a08d233cca19\") " pod="openshift-marketplace/redhat-marketplace-7bmzz" Dec 03 17:50:21 crc kubenswrapper[4998]: I1203 17:50:21.911681 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7bmzz" Dec 03 17:50:22 crc kubenswrapper[4998]: I1203 17:50:22.420156 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7bmzz"] Dec 03 17:50:23 crc kubenswrapper[4998]: I1203 17:50:23.048583 4998 generic.go:334] "Generic (PLEG): container finished" podID="9fe16aa3-7744-4763-869b-a08d233cca19" containerID="7e7e79c35b6b6cd929257324b4bb5e0e75bb3ba2fa69c1b5fc2ee37cdc3fac43" exitCode=0 Dec 03 17:50:23 crc kubenswrapper[4998]: I1203 17:50:23.048683 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7bmzz" event={"ID":"9fe16aa3-7744-4763-869b-a08d233cca19","Type":"ContainerDied","Data":"7e7e79c35b6b6cd929257324b4bb5e0e75bb3ba2fa69c1b5fc2ee37cdc3fac43"} Dec 03 17:50:23 crc kubenswrapper[4998]: I1203 17:50:23.049036 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7bmzz" event={"ID":"9fe16aa3-7744-4763-869b-a08d233cca19","Type":"ContainerStarted","Data":"5b38418ea5db116390f6ee8302795242d6ab80f32dc794b1f53495b5c374577b"} Dec 03 17:50:24 crc kubenswrapper[4998]: I1203 17:50:24.678675 4998 scope.go:117] "RemoveContainer" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" Dec 03 17:50:24 crc kubenswrapper[4998]: E1203 17:50:24.679538 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:50:25 crc kubenswrapper[4998]: I1203 17:50:25.104481 4998 generic.go:334] "Generic (PLEG): container finished" podID="9fe16aa3-7744-4763-869b-a08d233cca19" containerID="cb03d669f198e0f2a21e826eed9bb25734b57e809d65d4485de5d67dd992b0de" exitCode=0 Dec 03 17:50:25 crc kubenswrapper[4998]: I1203 17:50:25.104552 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7bmzz" event={"ID":"9fe16aa3-7744-4763-869b-a08d233cca19","Type":"ContainerDied","Data":"cb03d669f198e0f2a21e826eed9bb25734b57e809d65d4485de5d67dd992b0de"} Dec 03 17:50:26 crc kubenswrapper[4998]: I1203 17:50:26.115252 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7bmzz" event={"ID":"9fe16aa3-7744-4763-869b-a08d233cca19","Type":"ContainerStarted","Data":"5bf04ee47ec0652fb581dd966b81eeaa34c4fbe44c104dc17b8e685c8c8d952f"} Dec 03 17:50:26 crc kubenswrapper[4998]: I1203 17:50:26.149138 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7bmzz" podStartSLOduration=2.67699339 podStartE2EDuration="5.149119564s" podCreationTimestamp="2025-12-03 17:50:21 +0000 UTC" firstStartedPulling="2025-12-03 17:50:23.051911931 +0000 UTC m=+6401.663612154" lastFinishedPulling="2025-12-03 17:50:25.524038065 +0000 UTC m=+6404.135738328" observedRunningTime="2025-12-03 17:50:26.137349191 +0000 UTC m=+6404.749049414" watchObservedRunningTime="2025-12-03 17:50:26.149119564 +0000 UTC m=+6404.760819777" Dec 03 17:50:28 crc kubenswrapper[4998]: I1203 17:50:28.140544 4998 generic.go:334] "Generic (PLEG): container finished" podID="5f4e13e1-8408-4046-a8b4-7591e8a1bd22" containerID="156ed544489d596d38766112ab42a92f8d9d6cfa55075b8d2e59bad36d942f78" exitCode=0 Dec 03 17:50:28 crc kubenswrapper[4998]: I1203 17:50:28.140652 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jx2sk/must-gather-nbb85" event={"ID":"5f4e13e1-8408-4046-a8b4-7591e8a1bd22","Type":"ContainerDied","Data":"156ed544489d596d38766112ab42a92f8d9d6cfa55075b8d2e59bad36d942f78"} Dec 03 17:50:28 crc kubenswrapper[4998]: I1203 17:50:28.141839 4998 scope.go:117] "RemoveContainer" containerID="156ed544489d596d38766112ab42a92f8d9d6cfa55075b8d2e59bad36d942f78" Dec 03 17:50:29 crc kubenswrapper[4998]: I1203 17:50:29.147832 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jx2sk_must-gather-nbb85_5f4e13e1-8408-4046-a8b4-7591e8a1bd22/gather/0.log" Dec 03 17:50:31 crc kubenswrapper[4998]: I1203 17:50:31.912816 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7bmzz" Dec 03 17:50:31 crc kubenswrapper[4998]: I1203 17:50:31.913157 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7bmzz" Dec 03 17:50:31 crc kubenswrapper[4998]: I1203 17:50:31.973474 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7bmzz" Dec 03 17:50:32 crc kubenswrapper[4998]: I1203 17:50:32.254428 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7bmzz" Dec 03 17:50:32 crc kubenswrapper[4998]: I1203 17:50:32.304969 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7bmzz"] Dec 03 17:50:34 crc kubenswrapper[4998]: I1203 17:50:34.211651 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7bmzz" podUID="9fe16aa3-7744-4763-869b-a08d233cca19" containerName="registry-server" containerID="cri-o://5bf04ee47ec0652fb581dd966b81eeaa34c4fbe44c104dc17b8e685c8c8d952f" gracePeriod=2 Dec 03 17:50:34 crc kubenswrapper[4998]: I1203 17:50:34.742663 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7bmzz" Dec 03 17:50:34 crc kubenswrapper[4998]: I1203 17:50:34.868084 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87kxv\" (UniqueName: \"kubernetes.io/projected/9fe16aa3-7744-4763-869b-a08d233cca19-kube-api-access-87kxv\") pod \"9fe16aa3-7744-4763-869b-a08d233cca19\" (UID: \"9fe16aa3-7744-4763-869b-a08d233cca19\") " Dec 03 17:50:34 crc kubenswrapper[4998]: I1203 17:50:34.868481 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fe16aa3-7744-4763-869b-a08d233cca19-utilities\") pod \"9fe16aa3-7744-4763-869b-a08d233cca19\" (UID: \"9fe16aa3-7744-4763-869b-a08d233cca19\") " Dec 03 17:50:34 crc kubenswrapper[4998]: I1203 17:50:34.868676 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fe16aa3-7744-4763-869b-a08d233cca19-catalog-content\") pod \"9fe16aa3-7744-4763-869b-a08d233cca19\" (UID: \"9fe16aa3-7744-4763-869b-a08d233cca19\") " Dec 03 17:50:34 crc kubenswrapper[4998]: I1203 17:50:34.869819 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fe16aa3-7744-4763-869b-a08d233cca19-utilities" (OuterVolumeSpecName: "utilities") pod "9fe16aa3-7744-4763-869b-a08d233cca19" (UID: "9fe16aa3-7744-4763-869b-a08d233cca19"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:50:34 crc kubenswrapper[4998]: I1203 17:50:34.880958 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fe16aa3-7744-4763-869b-a08d233cca19-kube-api-access-87kxv" (OuterVolumeSpecName: "kube-api-access-87kxv") pod "9fe16aa3-7744-4763-869b-a08d233cca19" (UID: "9fe16aa3-7744-4763-869b-a08d233cca19"). InnerVolumeSpecName "kube-api-access-87kxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:50:34 crc kubenswrapper[4998]: I1203 17:50:34.888332 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fe16aa3-7744-4763-869b-a08d233cca19-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9fe16aa3-7744-4763-869b-a08d233cca19" (UID: "9fe16aa3-7744-4763-869b-a08d233cca19"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:50:34 crc kubenswrapper[4998]: I1203 17:50:34.971053 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87kxv\" (UniqueName: \"kubernetes.io/projected/9fe16aa3-7744-4763-869b-a08d233cca19-kube-api-access-87kxv\") on node \"crc\" DevicePath \"\"" Dec 03 17:50:34 crc kubenswrapper[4998]: I1203 17:50:34.971090 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fe16aa3-7744-4763-869b-a08d233cca19-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:50:34 crc kubenswrapper[4998]: I1203 17:50:34.971103 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fe16aa3-7744-4763-869b-a08d233cca19-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:50:35 crc kubenswrapper[4998]: I1203 17:50:35.228318 4998 generic.go:334] "Generic (PLEG): container finished" podID="9fe16aa3-7744-4763-869b-a08d233cca19" containerID="5bf04ee47ec0652fb581dd966b81eeaa34c4fbe44c104dc17b8e685c8c8d952f" exitCode=0 Dec 03 17:50:35 crc kubenswrapper[4998]: I1203 17:50:35.228491 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7bmzz" event={"ID":"9fe16aa3-7744-4763-869b-a08d233cca19","Type":"ContainerDied","Data":"5bf04ee47ec0652fb581dd966b81eeaa34c4fbe44c104dc17b8e685c8c8d952f"} Dec 03 17:50:35 crc kubenswrapper[4998]: I1203 17:50:35.228894 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7bmzz" event={"ID":"9fe16aa3-7744-4763-869b-a08d233cca19","Type":"ContainerDied","Data":"5b38418ea5db116390f6ee8302795242d6ab80f32dc794b1f53495b5c374577b"} Dec 03 17:50:35 crc kubenswrapper[4998]: I1203 17:50:35.228907 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7bmzz" Dec 03 17:50:35 crc kubenswrapper[4998]: I1203 17:50:35.228925 4998 scope.go:117] "RemoveContainer" containerID="5bf04ee47ec0652fb581dd966b81eeaa34c4fbe44c104dc17b8e685c8c8d952f" Dec 03 17:50:35 crc kubenswrapper[4998]: I1203 17:50:35.251315 4998 scope.go:117] "RemoveContainer" containerID="cb03d669f198e0f2a21e826eed9bb25734b57e809d65d4485de5d67dd992b0de" Dec 03 17:50:35 crc kubenswrapper[4998]: I1203 17:50:35.281868 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7bmzz"] Dec 03 17:50:35 crc kubenswrapper[4998]: I1203 17:50:35.284128 4998 scope.go:117] "RemoveContainer" containerID="7e7e79c35b6b6cd929257324b4bb5e0e75bb3ba2fa69c1b5fc2ee37cdc3fac43" Dec 03 17:50:35 crc kubenswrapper[4998]: I1203 17:50:35.299358 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7bmzz"] Dec 03 17:50:35 crc kubenswrapper[4998]: I1203 17:50:35.335398 4998 scope.go:117] "RemoveContainer" containerID="5bf04ee47ec0652fb581dd966b81eeaa34c4fbe44c104dc17b8e685c8c8d952f" Dec 03 17:50:35 crc kubenswrapper[4998]: E1203 17:50:35.335820 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bf04ee47ec0652fb581dd966b81eeaa34c4fbe44c104dc17b8e685c8c8d952f\": container with ID starting with 5bf04ee47ec0652fb581dd966b81eeaa34c4fbe44c104dc17b8e685c8c8d952f not found: ID does not exist" containerID="5bf04ee47ec0652fb581dd966b81eeaa34c4fbe44c104dc17b8e685c8c8d952f" Dec 03 17:50:35 crc kubenswrapper[4998]: I1203 17:50:35.335938 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bf04ee47ec0652fb581dd966b81eeaa34c4fbe44c104dc17b8e685c8c8d952f"} err="failed to get container status \"5bf04ee47ec0652fb581dd966b81eeaa34c4fbe44c104dc17b8e685c8c8d952f\": rpc error: code = NotFound desc = could not find container \"5bf04ee47ec0652fb581dd966b81eeaa34c4fbe44c104dc17b8e685c8c8d952f\": container with ID starting with 5bf04ee47ec0652fb581dd966b81eeaa34c4fbe44c104dc17b8e685c8c8d952f not found: ID does not exist" Dec 03 17:50:35 crc kubenswrapper[4998]: I1203 17:50:35.336021 4998 scope.go:117] "RemoveContainer" containerID="cb03d669f198e0f2a21e826eed9bb25734b57e809d65d4485de5d67dd992b0de" Dec 03 17:50:35 crc kubenswrapper[4998]: E1203 17:50:35.336562 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb03d669f198e0f2a21e826eed9bb25734b57e809d65d4485de5d67dd992b0de\": container with ID starting with cb03d669f198e0f2a21e826eed9bb25734b57e809d65d4485de5d67dd992b0de not found: ID does not exist" containerID="cb03d669f198e0f2a21e826eed9bb25734b57e809d65d4485de5d67dd992b0de" Dec 03 17:50:35 crc kubenswrapper[4998]: I1203 17:50:35.336649 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb03d669f198e0f2a21e826eed9bb25734b57e809d65d4485de5d67dd992b0de"} err="failed to get container status \"cb03d669f198e0f2a21e826eed9bb25734b57e809d65d4485de5d67dd992b0de\": rpc error: code = NotFound desc = could not find container \"cb03d669f198e0f2a21e826eed9bb25734b57e809d65d4485de5d67dd992b0de\": container with ID starting with cb03d669f198e0f2a21e826eed9bb25734b57e809d65d4485de5d67dd992b0de not found: ID does not exist" Dec 03 17:50:35 crc kubenswrapper[4998]: I1203 17:50:35.336718 4998 scope.go:117] "RemoveContainer" containerID="7e7e79c35b6b6cd929257324b4bb5e0e75bb3ba2fa69c1b5fc2ee37cdc3fac43" Dec 03 17:50:35 crc kubenswrapper[4998]: E1203 17:50:35.337095 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e7e79c35b6b6cd929257324b4bb5e0e75bb3ba2fa69c1b5fc2ee37cdc3fac43\": container with ID starting with 7e7e79c35b6b6cd929257324b4bb5e0e75bb3ba2fa69c1b5fc2ee37cdc3fac43 not found: ID does not exist" containerID="7e7e79c35b6b6cd929257324b4bb5e0e75bb3ba2fa69c1b5fc2ee37cdc3fac43" Dec 03 17:50:35 crc kubenswrapper[4998]: I1203 17:50:35.337174 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e7e79c35b6b6cd929257324b4bb5e0e75bb3ba2fa69c1b5fc2ee37cdc3fac43"} err="failed to get container status \"7e7e79c35b6b6cd929257324b4bb5e0e75bb3ba2fa69c1b5fc2ee37cdc3fac43\": rpc error: code = NotFound desc = could not find container \"7e7e79c35b6b6cd929257324b4bb5e0e75bb3ba2fa69c1b5fc2ee37cdc3fac43\": container with ID starting with 7e7e79c35b6b6cd929257324b4bb5e0e75bb3ba2fa69c1b5fc2ee37cdc3fac43 not found: ID does not exist" Dec 03 17:50:35 crc kubenswrapper[4998]: I1203 17:50:35.694179 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fe16aa3-7744-4763-869b-a08d233cca19" path="/var/lib/kubelet/pods/9fe16aa3-7744-4763-869b-a08d233cca19/volumes" Dec 03 17:50:37 crc kubenswrapper[4998]: I1203 17:50:37.342632 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jx2sk/must-gather-nbb85"] Dec 03 17:50:37 crc kubenswrapper[4998]: I1203 17:50:37.343417 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-jx2sk/must-gather-nbb85" podUID="5f4e13e1-8408-4046-a8b4-7591e8a1bd22" containerName="copy" containerID="cri-o://7615d5fcfd5c593a72110f7c17562e3cb44baf60e638c9f16434f81a459e543e" gracePeriod=2 Dec 03 17:50:37 crc kubenswrapper[4998]: I1203 17:50:37.360945 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jx2sk/must-gather-nbb85"] Dec 03 17:50:37 crc kubenswrapper[4998]: I1203 17:50:37.815396 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jx2sk_must-gather-nbb85_5f4e13e1-8408-4046-a8b4-7591e8a1bd22/copy/0.log" Dec 03 17:50:37 crc kubenswrapper[4998]: I1203 17:50:37.816150 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jx2sk/must-gather-nbb85" Dec 03 17:50:37 crc kubenswrapper[4998]: I1203 17:50:37.955381 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vh92\" (UniqueName: \"kubernetes.io/projected/5f4e13e1-8408-4046-a8b4-7591e8a1bd22-kube-api-access-7vh92\") pod \"5f4e13e1-8408-4046-a8b4-7591e8a1bd22\" (UID: \"5f4e13e1-8408-4046-a8b4-7591e8a1bd22\") " Dec 03 17:50:37 crc kubenswrapper[4998]: I1203 17:50:37.955663 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5f4e13e1-8408-4046-a8b4-7591e8a1bd22-must-gather-output\") pod \"5f4e13e1-8408-4046-a8b4-7591e8a1bd22\" (UID: \"5f4e13e1-8408-4046-a8b4-7591e8a1bd22\") " Dec 03 17:50:37 crc kubenswrapper[4998]: I1203 17:50:37.969222 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f4e13e1-8408-4046-a8b4-7591e8a1bd22-kube-api-access-7vh92" (OuterVolumeSpecName: "kube-api-access-7vh92") pod "5f4e13e1-8408-4046-a8b4-7591e8a1bd22" (UID: "5f4e13e1-8408-4046-a8b4-7591e8a1bd22"). InnerVolumeSpecName "kube-api-access-7vh92". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:50:38 crc kubenswrapper[4998]: I1203 17:50:38.058518 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vh92\" (UniqueName: \"kubernetes.io/projected/5f4e13e1-8408-4046-a8b4-7591e8a1bd22-kube-api-access-7vh92\") on node \"crc\" DevicePath \"\"" Dec 03 17:50:38 crc kubenswrapper[4998]: I1203 17:50:38.183797 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f4e13e1-8408-4046-a8b4-7591e8a1bd22-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "5f4e13e1-8408-4046-a8b4-7591e8a1bd22" (UID: "5f4e13e1-8408-4046-a8b4-7591e8a1bd22"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:50:38 crc kubenswrapper[4998]: I1203 17:50:38.266187 4998 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5f4e13e1-8408-4046-a8b4-7591e8a1bd22-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 03 17:50:38 crc kubenswrapper[4998]: I1203 17:50:38.269799 4998 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jx2sk_must-gather-nbb85_5f4e13e1-8408-4046-a8b4-7591e8a1bd22/copy/0.log" Dec 03 17:50:38 crc kubenswrapper[4998]: I1203 17:50:38.270354 4998 generic.go:334] "Generic (PLEG): container finished" podID="5f4e13e1-8408-4046-a8b4-7591e8a1bd22" containerID="7615d5fcfd5c593a72110f7c17562e3cb44baf60e638c9f16434f81a459e543e" exitCode=143 Dec 03 17:50:38 crc kubenswrapper[4998]: I1203 17:50:38.270421 4998 scope.go:117] "RemoveContainer" containerID="7615d5fcfd5c593a72110f7c17562e3cb44baf60e638c9f16434f81a459e543e" Dec 03 17:50:38 crc kubenswrapper[4998]: I1203 17:50:38.270434 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jx2sk/must-gather-nbb85" Dec 03 17:50:38 crc kubenswrapper[4998]: I1203 17:50:38.323960 4998 scope.go:117] "RemoveContainer" containerID="156ed544489d596d38766112ab42a92f8d9d6cfa55075b8d2e59bad36d942f78" Dec 03 17:50:38 crc kubenswrapper[4998]: I1203 17:50:38.391250 4998 scope.go:117] "RemoveContainer" containerID="7615d5fcfd5c593a72110f7c17562e3cb44baf60e638c9f16434f81a459e543e" Dec 03 17:50:38 crc kubenswrapper[4998]: E1203 17:50:38.395668 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7615d5fcfd5c593a72110f7c17562e3cb44baf60e638c9f16434f81a459e543e\": container with ID starting with 7615d5fcfd5c593a72110f7c17562e3cb44baf60e638c9f16434f81a459e543e not found: ID does not exist" containerID="7615d5fcfd5c593a72110f7c17562e3cb44baf60e638c9f16434f81a459e543e" Dec 03 17:50:38 crc kubenswrapper[4998]: I1203 17:50:38.395707 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7615d5fcfd5c593a72110f7c17562e3cb44baf60e638c9f16434f81a459e543e"} err="failed to get container status \"7615d5fcfd5c593a72110f7c17562e3cb44baf60e638c9f16434f81a459e543e\": rpc error: code = NotFound desc = could not find container \"7615d5fcfd5c593a72110f7c17562e3cb44baf60e638c9f16434f81a459e543e\": container with ID starting with 7615d5fcfd5c593a72110f7c17562e3cb44baf60e638c9f16434f81a459e543e not found: ID does not exist" Dec 03 17:50:38 crc kubenswrapper[4998]: I1203 17:50:38.395731 4998 scope.go:117] "RemoveContainer" containerID="156ed544489d596d38766112ab42a92f8d9d6cfa55075b8d2e59bad36d942f78" Dec 03 17:50:38 crc kubenswrapper[4998]: E1203 17:50:38.401725 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"156ed544489d596d38766112ab42a92f8d9d6cfa55075b8d2e59bad36d942f78\": container with ID starting with 156ed544489d596d38766112ab42a92f8d9d6cfa55075b8d2e59bad36d942f78 not found: ID does not exist" containerID="156ed544489d596d38766112ab42a92f8d9d6cfa55075b8d2e59bad36d942f78" Dec 03 17:50:38 crc kubenswrapper[4998]: I1203 17:50:38.401924 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"156ed544489d596d38766112ab42a92f8d9d6cfa55075b8d2e59bad36d942f78"} err="failed to get container status \"156ed544489d596d38766112ab42a92f8d9d6cfa55075b8d2e59bad36d942f78\": rpc error: code = NotFound desc = could not find container \"156ed544489d596d38766112ab42a92f8d9d6cfa55075b8d2e59bad36d942f78\": container with ID starting with 156ed544489d596d38766112ab42a92f8d9d6cfa55075b8d2e59bad36d942f78 not found: ID does not exist" Dec 03 17:50:39 crc kubenswrapper[4998]: I1203 17:50:39.678171 4998 scope.go:117] "RemoveContainer" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" Dec 03 17:50:39 crc kubenswrapper[4998]: E1203 17:50:39.679221 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:50:39 crc kubenswrapper[4998]: I1203 17:50:39.695595 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f4e13e1-8408-4046-a8b4-7591e8a1bd22" path="/var/lib/kubelet/pods/5f4e13e1-8408-4046-a8b4-7591e8a1bd22/volumes" Dec 03 17:50:48 crc kubenswrapper[4998]: I1203 17:50:48.947964 4998 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-69brf"] Dec 03 17:50:48 crc kubenswrapper[4998]: E1203 17:50:48.949274 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f4e13e1-8408-4046-a8b4-7591e8a1bd22" containerName="gather" Dec 03 17:50:48 crc kubenswrapper[4998]: I1203 17:50:48.949297 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f4e13e1-8408-4046-a8b4-7591e8a1bd22" containerName="gather" Dec 03 17:50:48 crc kubenswrapper[4998]: E1203 17:50:48.949325 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fe16aa3-7744-4763-869b-a08d233cca19" containerName="extract-utilities" Dec 03 17:50:48 crc kubenswrapper[4998]: I1203 17:50:48.949337 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fe16aa3-7744-4763-869b-a08d233cca19" containerName="extract-utilities" Dec 03 17:50:48 crc kubenswrapper[4998]: E1203 17:50:48.949383 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fe16aa3-7744-4763-869b-a08d233cca19" containerName="registry-server" Dec 03 17:50:48 crc kubenswrapper[4998]: I1203 17:50:48.949395 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fe16aa3-7744-4763-869b-a08d233cca19" containerName="registry-server" Dec 03 17:50:48 crc kubenswrapper[4998]: E1203 17:50:48.949408 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fe16aa3-7744-4763-869b-a08d233cca19" containerName="extract-content" Dec 03 17:50:48 crc kubenswrapper[4998]: I1203 17:50:48.949420 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fe16aa3-7744-4763-869b-a08d233cca19" containerName="extract-content" Dec 03 17:50:48 crc kubenswrapper[4998]: E1203 17:50:48.949445 4998 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f4e13e1-8408-4046-a8b4-7591e8a1bd22" containerName="copy" Dec 03 17:50:48 crc kubenswrapper[4998]: I1203 17:50:48.949456 4998 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f4e13e1-8408-4046-a8b4-7591e8a1bd22" containerName="copy" Dec 03 17:50:48 crc kubenswrapper[4998]: I1203 17:50:48.949846 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f4e13e1-8408-4046-a8b4-7591e8a1bd22" containerName="gather" Dec 03 17:50:48 crc kubenswrapper[4998]: I1203 17:50:48.949895 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f4e13e1-8408-4046-a8b4-7591e8a1bd22" containerName="copy" Dec 03 17:50:48 crc kubenswrapper[4998]: I1203 17:50:48.949913 4998 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fe16aa3-7744-4763-869b-a08d233cca19" containerName="registry-server" Dec 03 17:50:48 crc kubenswrapper[4998]: I1203 17:50:48.952587 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-69brf" Dec 03 17:50:48 crc kubenswrapper[4998]: I1203 17:50:48.960358 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-69brf"] Dec 03 17:50:49 crc kubenswrapper[4998]: I1203 17:50:49.128901 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpzzz\" (UniqueName: \"kubernetes.io/projected/546f8c42-b5c3-4e54-8841-dd5711861cdc-kube-api-access-hpzzz\") pod \"community-operators-69brf\" (UID: \"546f8c42-b5c3-4e54-8841-dd5711861cdc\") " pod="openshift-marketplace/community-operators-69brf" Dec 03 17:50:49 crc kubenswrapper[4998]: I1203 17:50:49.128963 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/546f8c42-b5c3-4e54-8841-dd5711861cdc-utilities\") pod \"community-operators-69brf\" (UID: \"546f8c42-b5c3-4e54-8841-dd5711861cdc\") " pod="openshift-marketplace/community-operators-69brf" Dec 03 17:50:49 crc kubenswrapper[4998]: I1203 17:50:49.129037 4998 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/546f8c42-b5c3-4e54-8841-dd5711861cdc-catalog-content\") pod \"community-operators-69brf\" (UID: \"546f8c42-b5c3-4e54-8841-dd5711861cdc\") " pod="openshift-marketplace/community-operators-69brf" Dec 03 17:50:49 crc kubenswrapper[4998]: I1203 17:50:49.230809 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/546f8c42-b5c3-4e54-8841-dd5711861cdc-catalog-content\") pod \"community-operators-69brf\" (UID: \"546f8c42-b5c3-4e54-8841-dd5711861cdc\") " pod="openshift-marketplace/community-operators-69brf" Dec 03 17:50:49 crc kubenswrapper[4998]: I1203 17:50:49.230995 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpzzz\" (UniqueName: \"kubernetes.io/projected/546f8c42-b5c3-4e54-8841-dd5711861cdc-kube-api-access-hpzzz\") pod \"community-operators-69brf\" (UID: \"546f8c42-b5c3-4e54-8841-dd5711861cdc\") " pod="openshift-marketplace/community-operators-69brf" Dec 03 17:50:49 crc kubenswrapper[4998]: I1203 17:50:49.231027 4998 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/546f8c42-b5c3-4e54-8841-dd5711861cdc-utilities\") pod \"community-operators-69brf\" (UID: \"546f8c42-b5c3-4e54-8841-dd5711861cdc\") " pod="openshift-marketplace/community-operators-69brf" Dec 03 17:50:49 crc kubenswrapper[4998]: I1203 17:50:49.231445 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/546f8c42-b5c3-4e54-8841-dd5711861cdc-catalog-content\") pod \"community-operators-69brf\" (UID: \"546f8c42-b5c3-4e54-8841-dd5711861cdc\") " pod="openshift-marketplace/community-operators-69brf" Dec 03 17:50:49 crc kubenswrapper[4998]: I1203 17:50:49.231515 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/546f8c42-b5c3-4e54-8841-dd5711861cdc-utilities\") pod \"community-operators-69brf\" (UID: \"546f8c42-b5c3-4e54-8841-dd5711861cdc\") " pod="openshift-marketplace/community-operators-69brf" Dec 03 17:50:49 crc kubenswrapper[4998]: I1203 17:50:49.249584 4998 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpzzz\" (UniqueName: \"kubernetes.io/projected/546f8c42-b5c3-4e54-8841-dd5711861cdc-kube-api-access-hpzzz\") pod \"community-operators-69brf\" (UID: \"546f8c42-b5c3-4e54-8841-dd5711861cdc\") " pod="openshift-marketplace/community-operators-69brf" Dec 03 17:50:49 crc kubenswrapper[4998]: I1203 17:50:49.304482 4998 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-69brf" Dec 03 17:50:49 crc kubenswrapper[4998]: I1203 17:50:49.824655 4998 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-69brf"] Dec 03 17:50:50 crc kubenswrapper[4998]: I1203 17:50:50.438984 4998 generic.go:334] "Generic (PLEG): container finished" podID="546f8c42-b5c3-4e54-8841-dd5711861cdc" containerID="b4dd1c1b05b9f8f129376ece9fbb152a551595eb076c0878c30fbea5f8e548c9" exitCode=0 Dec 03 17:50:50 crc kubenswrapper[4998]: I1203 17:50:50.439047 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-69brf" event={"ID":"546f8c42-b5c3-4e54-8841-dd5711861cdc","Type":"ContainerDied","Data":"b4dd1c1b05b9f8f129376ece9fbb152a551595eb076c0878c30fbea5f8e548c9"} Dec 03 17:50:50 crc kubenswrapper[4998]: I1203 17:50:50.439306 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-69brf" event={"ID":"546f8c42-b5c3-4e54-8841-dd5711861cdc","Type":"ContainerStarted","Data":"d771ce7277bc778e85278624d2209494d9f86af4819ff510a874ff97b5fc41f8"} Dec 03 17:50:50 crc kubenswrapper[4998]: I1203 17:50:50.442222 4998 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 17:50:50 crc kubenswrapper[4998]: I1203 17:50:50.681039 4998 scope.go:117] "RemoveContainer" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" Dec 03 17:50:50 crc kubenswrapper[4998]: E1203 17:50:50.681497 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:50:52 crc kubenswrapper[4998]: I1203 17:50:52.464688 4998 generic.go:334] "Generic (PLEG): container finished" podID="546f8c42-b5c3-4e54-8841-dd5711861cdc" containerID="30d1fcb0d100251fd50263a1d46c0ec745bbcf604f4bd4671198608dee27b077" exitCode=0 Dec 03 17:50:52 crc kubenswrapper[4998]: I1203 17:50:52.464806 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-69brf" event={"ID":"546f8c42-b5c3-4e54-8841-dd5711861cdc","Type":"ContainerDied","Data":"30d1fcb0d100251fd50263a1d46c0ec745bbcf604f4bd4671198608dee27b077"} Dec 03 17:50:53 crc kubenswrapper[4998]: I1203 17:50:53.479917 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-69brf" event={"ID":"546f8c42-b5c3-4e54-8841-dd5711861cdc","Type":"ContainerStarted","Data":"38f745a141882cb83bf273a9705d9836aabb8abad08540b3c702f0c62ba1b7e2"} Dec 03 17:50:53 crc kubenswrapper[4998]: I1203 17:50:53.504517 4998 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-69brf" podStartSLOduration=3.057982916 podStartE2EDuration="5.504492664s" podCreationTimestamp="2025-12-03 17:50:48 +0000 UTC" firstStartedPulling="2025-12-03 17:50:50.44182318 +0000 UTC m=+6429.053523433" lastFinishedPulling="2025-12-03 17:50:52.888332948 +0000 UTC m=+6431.500033181" observedRunningTime="2025-12-03 17:50:53.497523931 +0000 UTC m=+6432.109224174" watchObservedRunningTime="2025-12-03 17:50:53.504492664 +0000 UTC m=+6432.116192917" Dec 03 17:50:59 crc kubenswrapper[4998]: I1203 17:50:59.305923 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-69brf" Dec 03 17:50:59 crc kubenswrapper[4998]: I1203 17:50:59.306576 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-69brf" Dec 03 17:50:59 crc kubenswrapper[4998]: I1203 17:50:59.399946 4998 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-69brf" Dec 03 17:50:59 crc kubenswrapper[4998]: I1203 17:50:59.690107 4998 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-69brf" Dec 03 17:50:59 crc kubenswrapper[4998]: I1203 17:50:59.766838 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-69brf"] Dec 03 17:51:01 crc kubenswrapper[4998]: I1203 17:51:01.582520 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-69brf" podUID="546f8c42-b5c3-4e54-8841-dd5711861cdc" containerName="registry-server" containerID="cri-o://38f745a141882cb83bf273a9705d9836aabb8abad08540b3c702f0c62ba1b7e2" gracePeriod=2 Dec 03 17:51:01 crc kubenswrapper[4998]: I1203 17:51:01.685679 4998 scope.go:117] "RemoveContainer" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" Dec 03 17:51:01 crc kubenswrapper[4998]: E1203 17:51:01.688148 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:51:02 crc kubenswrapper[4998]: I1203 17:51:02.142807 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-69brf" Dec 03 17:51:02 crc kubenswrapper[4998]: I1203 17:51:02.150618 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpzzz\" (UniqueName: \"kubernetes.io/projected/546f8c42-b5c3-4e54-8841-dd5711861cdc-kube-api-access-hpzzz\") pod \"546f8c42-b5c3-4e54-8841-dd5711861cdc\" (UID: \"546f8c42-b5c3-4e54-8841-dd5711861cdc\") " Dec 03 17:51:02 crc kubenswrapper[4998]: I1203 17:51:02.150720 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/546f8c42-b5c3-4e54-8841-dd5711861cdc-utilities\") pod \"546f8c42-b5c3-4e54-8841-dd5711861cdc\" (UID: \"546f8c42-b5c3-4e54-8841-dd5711861cdc\") " Dec 03 17:51:02 crc kubenswrapper[4998]: I1203 17:51:02.150866 4998 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/546f8c42-b5c3-4e54-8841-dd5711861cdc-catalog-content\") pod \"546f8c42-b5c3-4e54-8841-dd5711861cdc\" (UID: \"546f8c42-b5c3-4e54-8841-dd5711861cdc\") " Dec 03 17:51:02 crc kubenswrapper[4998]: I1203 17:51:02.151749 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/546f8c42-b5c3-4e54-8841-dd5711861cdc-utilities" (OuterVolumeSpecName: "utilities") pod "546f8c42-b5c3-4e54-8841-dd5711861cdc" (UID: "546f8c42-b5c3-4e54-8841-dd5711861cdc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:51:02 crc kubenswrapper[4998]: I1203 17:51:02.152016 4998 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/546f8c42-b5c3-4e54-8841-dd5711861cdc-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:51:02 crc kubenswrapper[4998]: I1203 17:51:02.161610 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/546f8c42-b5c3-4e54-8841-dd5711861cdc-kube-api-access-hpzzz" (OuterVolumeSpecName: "kube-api-access-hpzzz") pod "546f8c42-b5c3-4e54-8841-dd5711861cdc" (UID: "546f8c42-b5c3-4e54-8841-dd5711861cdc"). InnerVolumeSpecName "kube-api-access-hpzzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:51:02 crc kubenswrapper[4998]: I1203 17:51:02.208356 4998 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/546f8c42-b5c3-4e54-8841-dd5711861cdc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "546f8c42-b5c3-4e54-8841-dd5711861cdc" (UID: "546f8c42-b5c3-4e54-8841-dd5711861cdc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:51:02 crc kubenswrapper[4998]: I1203 17:51:02.254922 4998 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/546f8c42-b5c3-4e54-8841-dd5711861cdc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:51:02 crc kubenswrapper[4998]: I1203 17:51:02.255002 4998 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpzzz\" (UniqueName: \"kubernetes.io/projected/546f8c42-b5c3-4e54-8841-dd5711861cdc-kube-api-access-hpzzz\") on node \"crc\" DevicePath \"\"" Dec 03 17:51:02 crc kubenswrapper[4998]: I1203 17:51:02.603145 4998 generic.go:334] "Generic (PLEG): container finished" podID="546f8c42-b5c3-4e54-8841-dd5711861cdc" containerID="38f745a141882cb83bf273a9705d9836aabb8abad08540b3c702f0c62ba1b7e2" exitCode=0 Dec 03 17:51:02 crc kubenswrapper[4998]: I1203 17:51:02.603206 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-69brf" event={"ID":"546f8c42-b5c3-4e54-8841-dd5711861cdc","Type":"ContainerDied","Data":"38f745a141882cb83bf273a9705d9836aabb8abad08540b3c702f0c62ba1b7e2"} Dec 03 17:51:02 crc kubenswrapper[4998]: I1203 17:51:02.603249 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-69brf" event={"ID":"546f8c42-b5c3-4e54-8841-dd5711861cdc","Type":"ContainerDied","Data":"d771ce7277bc778e85278624d2209494d9f86af4819ff510a874ff97b5fc41f8"} Dec 03 17:51:02 crc kubenswrapper[4998]: I1203 17:51:02.603270 4998 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-69brf" Dec 03 17:51:02 crc kubenswrapper[4998]: I1203 17:51:02.603276 4998 scope.go:117] "RemoveContainer" containerID="38f745a141882cb83bf273a9705d9836aabb8abad08540b3c702f0c62ba1b7e2" Dec 03 17:51:02 crc kubenswrapper[4998]: I1203 17:51:02.655589 4998 scope.go:117] "RemoveContainer" containerID="30d1fcb0d100251fd50263a1d46c0ec745bbcf604f4bd4671198608dee27b077" Dec 03 17:51:02 crc kubenswrapper[4998]: I1203 17:51:02.660885 4998 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-69brf"] Dec 03 17:51:02 crc kubenswrapper[4998]: I1203 17:51:02.674269 4998 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-69brf"] Dec 03 17:51:02 crc kubenswrapper[4998]: I1203 17:51:02.683138 4998 scope.go:117] "RemoveContainer" containerID="b4dd1c1b05b9f8f129376ece9fbb152a551595eb076c0878c30fbea5f8e548c9" Dec 03 17:51:02 crc kubenswrapper[4998]: I1203 17:51:02.768158 4998 scope.go:117] "RemoveContainer" containerID="38f745a141882cb83bf273a9705d9836aabb8abad08540b3c702f0c62ba1b7e2" Dec 03 17:51:02 crc kubenswrapper[4998]: E1203 17:51:02.768902 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38f745a141882cb83bf273a9705d9836aabb8abad08540b3c702f0c62ba1b7e2\": container with ID starting with 38f745a141882cb83bf273a9705d9836aabb8abad08540b3c702f0c62ba1b7e2 not found: ID does not exist" containerID="38f745a141882cb83bf273a9705d9836aabb8abad08540b3c702f0c62ba1b7e2" Dec 03 17:51:02 crc kubenswrapper[4998]: I1203 17:51:02.769015 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38f745a141882cb83bf273a9705d9836aabb8abad08540b3c702f0c62ba1b7e2"} err="failed to get container status \"38f745a141882cb83bf273a9705d9836aabb8abad08540b3c702f0c62ba1b7e2\": rpc error: code = NotFound desc = could not find container \"38f745a141882cb83bf273a9705d9836aabb8abad08540b3c702f0c62ba1b7e2\": container with ID starting with 38f745a141882cb83bf273a9705d9836aabb8abad08540b3c702f0c62ba1b7e2 not found: ID does not exist" Dec 03 17:51:02 crc kubenswrapper[4998]: I1203 17:51:02.769069 4998 scope.go:117] "RemoveContainer" containerID="30d1fcb0d100251fd50263a1d46c0ec745bbcf604f4bd4671198608dee27b077" Dec 03 17:51:02 crc kubenswrapper[4998]: E1203 17:51:02.769505 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30d1fcb0d100251fd50263a1d46c0ec745bbcf604f4bd4671198608dee27b077\": container with ID starting with 30d1fcb0d100251fd50263a1d46c0ec745bbcf604f4bd4671198608dee27b077 not found: ID does not exist" containerID="30d1fcb0d100251fd50263a1d46c0ec745bbcf604f4bd4671198608dee27b077" Dec 03 17:51:02 crc kubenswrapper[4998]: I1203 17:51:02.769584 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30d1fcb0d100251fd50263a1d46c0ec745bbcf604f4bd4671198608dee27b077"} err="failed to get container status \"30d1fcb0d100251fd50263a1d46c0ec745bbcf604f4bd4671198608dee27b077\": rpc error: code = NotFound desc = could not find container \"30d1fcb0d100251fd50263a1d46c0ec745bbcf604f4bd4671198608dee27b077\": container with ID starting with 30d1fcb0d100251fd50263a1d46c0ec745bbcf604f4bd4671198608dee27b077 not found: ID does not exist" Dec 03 17:51:02 crc kubenswrapper[4998]: I1203 17:51:02.769630 4998 scope.go:117] "RemoveContainer" containerID="b4dd1c1b05b9f8f129376ece9fbb152a551595eb076c0878c30fbea5f8e548c9" Dec 03 17:51:02 crc kubenswrapper[4998]: E1203 17:51:02.770117 4998 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4dd1c1b05b9f8f129376ece9fbb152a551595eb076c0878c30fbea5f8e548c9\": container with ID starting with b4dd1c1b05b9f8f129376ece9fbb152a551595eb076c0878c30fbea5f8e548c9 not found: ID does not exist" containerID="b4dd1c1b05b9f8f129376ece9fbb152a551595eb076c0878c30fbea5f8e548c9" Dec 03 17:51:02 crc kubenswrapper[4998]: I1203 17:51:02.770187 4998 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4dd1c1b05b9f8f129376ece9fbb152a551595eb076c0878c30fbea5f8e548c9"} err="failed to get container status \"b4dd1c1b05b9f8f129376ece9fbb152a551595eb076c0878c30fbea5f8e548c9\": rpc error: code = NotFound desc = could not find container \"b4dd1c1b05b9f8f129376ece9fbb152a551595eb076c0878c30fbea5f8e548c9\": container with ID starting with b4dd1c1b05b9f8f129376ece9fbb152a551595eb076c0878c30fbea5f8e548c9 not found: ID does not exist" Dec 03 17:51:03 crc kubenswrapper[4998]: I1203 17:51:03.691686 4998 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="546f8c42-b5c3-4e54-8841-dd5711861cdc" path="/var/lib/kubelet/pods/546f8c42-b5c3-4e54-8841-dd5711861cdc/volumes" Dec 03 17:51:15 crc kubenswrapper[4998]: I1203 17:51:15.678451 4998 scope.go:117] "RemoveContainer" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" Dec 03 17:51:15 crc kubenswrapper[4998]: E1203 17:51:15.679906 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:51:28 crc kubenswrapper[4998]: I1203 17:51:28.678610 4998 scope.go:117] "RemoveContainer" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" Dec 03 17:51:28 crc kubenswrapper[4998]: E1203 17:51:28.679807 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:51:40 crc kubenswrapper[4998]: I1203 17:51:40.678163 4998 scope.go:117] "RemoveContainer" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" Dec 03 17:51:40 crc kubenswrapper[4998]: E1203 17:51:40.679385 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:51:54 crc kubenswrapper[4998]: I1203 17:51:54.678329 4998 scope.go:117] "RemoveContainer" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" Dec 03 17:51:54 crc kubenswrapper[4998]: E1203 17:51:54.679158 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:52:03 crc kubenswrapper[4998]: I1203 17:52:03.648905 4998 scope.go:117] "RemoveContainer" containerID="c2c1eb09b44eaa1c8b68aeab5f3c6629419e507a86be42dd66ce687350f49d9d" Dec 03 17:52:09 crc kubenswrapper[4998]: I1203 17:52:09.678348 4998 scope.go:117] "RemoveContainer" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" Dec 03 17:52:09 crc kubenswrapper[4998]: E1203 17:52:09.679577 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:52:22 crc kubenswrapper[4998]: I1203 17:52:22.678000 4998 scope.go:117] "RemoveContainer" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" Dec 03 17:52:22 crc kubenswrapper[4998]: E1203 17:52:22.678598 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:52:37 crc kubenswrapper[4998]: I1203 17:52:37.678799 4998 scope.go:117] "RemoveContainer" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" Dec 03 17:52:37 crc kubenswrapper[4998]: E1203 17:52:37.680121 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:52:49 crc kubenswrapper[4998]: I1203 17:52:49.678871 4998 scope.go:117] "RemoveContainer" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" Dec 03 17:52:49 crc kubenswrapper[4998]: E1203 17:52:49.679793 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:53:02 crc kubenswrapper[4998]: I1203 17:53:02.677889 4998 scope.go:117] "RemoveContainer" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" Dec 03 17:53:02 crc kubenswrapper[4998]: E1203 17:53:02.678973 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:53:15 crc kubenswrapper[4998]: I1203 17:53:15.693528 4998 scope.go:117] "RemoveContainer" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" Dec 03 17:53:15 crc kubenswrapper[4998]: E1203 17:53:15.694999 4998 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9fjbr_openshift-machine-config-operator(27aeb142-d4e3-4827-ac1f-c34de6822b14)\"" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" Dec 03 17:53:27 crc kubenswrapper[4998]: I1203 17:53:27.679259 4998 scope.go:117] "RemoveContainer" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" Dec 03 17:53:28 crc kubenswrapper[4998]: I1203 17:53:28.742798 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"92a4245f115a5ed315b78acd9aa744c2300f222a74616fdcf6f59cb1b66ff415"} Dec 03 17:55:27 crc kubenswrapper[4998]: I1203 17:55:27.111428 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:55:27 crc kubenswrapper[4998]: I1203 17:55:27.112254 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:55:57 crc kubenswrapper[4998]: I1203 17:55:57.111490 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:55:57 crc kubenswrapper[4998]: I1203 17:55:57.112258 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:56:27 crc kubenswrapper[4998]: I1203 17:56:27.111016 4998 patch_prober.go:28] interesting pod/machine-config-daemon-9fjbr container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:56:27 crc kubenswrapper[4998]: I1203 17:56:27.111680 4998 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:56:27 crc kubenswrapper[4998]: I1203 17:56:27.111749 4998 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" Dec 03 17:56:27 crc kubenswrapper[4998]: I1203 17:56:27.112946 4998 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"92a4245f115a5ed315b78acd9aa744c2300f222a74616fdcf6f59cb1b66ff415"} pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:56:27 crc kubenswrapper[4998]: I1203 17:56:27.113050 4998 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" podUID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerName="machine-config-daemon" containerID="cri-o://92a4245f115a5ed315b78acd9aa744c2300f222a74616fdcf6f59cb1b66ff415" gracePeriod=600 Dec 03 17:56:27 crc kubenswrapper[4998]: I1203 17:56:27.931273 4998 generic.go:334] "Generic (PLEG): container finished" podID="27aeb142-d4e3-4827-ac1f-c34de6822b14" containerID="92a4245f115a5ed315b78acd9aa744c2300f222a74616fdcf6f59cb1b66ff415" exitCode=0 Dec 03 17:56:27 crc kubenswrapper[4998]: I1203 17:56:27.931840 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerDied","Data":"92a4245f115a5ed315b78acd9aa744c2300f222a74616fdcf6f59cb1b66ff415"} Dec 03 17:56:27 crc kubenswrapper[4998]: I1203 17:56:27.931870 4998 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9fjbr" event={"ID":"27aeb142-d4e3-4827-ac1f-c34de6822b14","Type":"ContainerStarted","Data":"56b0fb9946ce3081c1c503b72a550d02888fd462a86f3912b44efda0f6d10e5d"} Dec 03 17:56:27 crc kubenswrapper[4998]: I1203 17:56:27.931888 4998 scope.go:117] "RemoveContainer" containerID="09b8613573107f093dd435e50b72b6fa39080bd9e4e50819365dca26e1336315" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114074745024455 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114074745017372 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114057163016510 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114057164015461 5ustar corecore